update_segment_keyword_index_task.py 2.7 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081
  1. import datetime
  2. import logging
  3. import time
  4. from typing import List, Optional
  5. import click
  6. from celery import shared_task
  7. from langchain.schema import Document
  8. from werkzeug.exceptions import NotFound
  9. from core.index.index import IndexBuilder
  10. from extensions.ext_database import db
  11. from extensions.ext_redis import redis_client
  12. from models.dataset import DocumentSegment
  13. @shared_task(queue='dataset')
  14. def update_segment_keyword_index_task(segment_id: str):
  15. """
  16. Async update segment index
  17. :param segment_id:
  18. Usage: update_segment_keyword_index_task.delay(segment_id)
  19. """
  20. logging.info(click.style('Start update segment keyword index: {}'.format(segment_id), fg='green'))
  21. start_at = time.perf_counter()
  22. segment = db.session.query(DocumentSegment).filter(DocumentSegment.id == segment_id).first()
  23. if not segment:
  24. raise NotFound('Segment not found')
  25. indexing_cache_key = 'segment_{}_indexing'.format(segment.id)
  26. try:
  27. dataset = segment.dataset
  28. if not dataset:
  29. logging.info(click.style('Segment {} has no dataset, pass.'.format(segment.id), fg='cyan'))
  30. return
  31. dataset_document = segment.document
  32. if not dataset_document:
  33. logging.info(click.style('Segment {} has no document, pass.'.format(segment.id), fg='cyan'))
  34. return
  35. if not dataset_document.enabled or dataset_document.archived or dataset_document.indexing_status != 'completed':
  36. logging.info(click.style('Segment {} document status is invalid, pass.'.format(segment.id), fg='cyan'))
  37. return
  38. kw_index = IndexBuilder.get_index(dataset, 'economy')
  39. # delete from keyword index
  40. kw_index.delete_by_ids([segment.index_node_id])
  41. # add new index
  42. document = Document(
  43. page_content=segment.content,
  44. metadata={
  45. "doc_id": segment.index_node_id,
  46. "doc_hash": segment.index_node_hash,
  47. "document_id": segment.document_id,
  48. "dataset_id": segment.dataset_id,
  49. }
  50. )
  51. # save keyword index
  52. index = IndexBuilder.get_index(dataset, 'economy')
  53. if index:
  54. index.update_segment_keywords_index(segment.index_node_id, segment.keywords)
  55. end_at = time.perf_counter()
  56. logging.info(click.style('Segment update index: {} latency: {}'.format(segment.id, end_at - start_at), fg='green'))
  57. except Exception as e:
  58. logging.exception("update segment index failed")
  59. segment.enabled = False
  60. segment.disabled_at = datetime.datetime.utcnow()
  61. segment.status = 'error'
  62. segment.error = str(e)
  63. db.session.commit()
  64. finally:
  65. redis_client.delete(indexing_cache_key)