enable_segment_to_index_task.py 2.7 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677
  1. import datetime
  2. import logging
  3. import time
  4. import click
  5. from celery import shared_task
  6. from werkzeug.exceptions import NotFound
  7. from core.rag.index_processor.index_processor_factory import IndexProcessorFactory
  8. from core.rag.models.document import Document
  9. from extensions.ext_database import db
  10. from extensions.ext_redis import redis_client
  11. from models.dataset import DocumentSegment
  12. @shared_task(queue='dataset')
  13. def enable_segment_to_index_task(segment_id: str):
  14. """
  15. Async enable segment to index
  16. :param segment_id:
  17. Usage: enable_segment_to_index_task.delay(segment_id)
  18. """
  19. logging.info(click.style('Start enable segment to index: {}'.format(segment_id), fg='green'))
  20. start_at = time.perf_counter()
  21. segment = db.session.query(DocumentSegment).filter(DocumentSegment.id == segment_id).first()
  22. if not segment:
  23. raise NotFound('Segment not found')
  24. if segment.status != 'completed':
  25. raise NotFound('Segment is not completed, enable action is not allowed.')
  26. indexing_cache_key = 'segment_{}_indexing'.format(segment.id)
  27. try:
  28. document = Document(
  29. page_content=segment.content,
  30. metadata={
  31. "doc_id": segment.index_node_id,
  32. "doc_hash": segment.index_node_hash,
  33. "document_id": segment.document_id,
  34. "dataset_id": segment.dataset_id,
  35. }
  36. )
  37. dataset = segment.dataset
  38. if not dataset:
  39. logging.info(click.style('Segment {} has no dataset, pass.'.format(segment.id), fg='cyan'))
  40. return
  41. dataset_document = segment.document
  42. if not dataset_document:
  43. logging.info(click.style('Segment {} has no document, pass.'.format(segment.id), fg='cyan'))
  44. return
  45. if not dataset_document.enabled or dataset_document.archived or dataset_document.indexing_status != 'completed':
  46. logging.info(click.style('Segment {} document status is invalid, pass.'.format(segment.id), fg='cyan'))
  47. return
  48. index_processor = IndexProcessorFactory(dataset_document.doc_form).init_index_processor()
  49. # save vector index
  50. index_processor.load(dataset, [document])
  51. end_at = time.perf_counter()
  52. logging.info(click.style('Segment enabled to index: {} latency: {}'.format(segment.id, end_at - start_at), fg='green'))
  53. except Exception as e:
  54. logging.exception("enable segment to index failed")
  55. segment.enabled = False
  56. segment.disabled_at = datetime.datetime.utcnow()
  57. segment.status = 'error'
  58. segment.error = str(e)
  59. db.session.commit()
  60. finally:
  61. redis_client.delete(indexing_cache_key)