clean_document_task.py 1.8 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152
  1. import logging
  2. import time
  3. import click
  4. from celery import shared_task
  5. from core.index.keyword_table_index import KeywordTableIndex
  6. from core.index.vector_index import VectorIndex
  7. from extensions.ext_database import db
  8. from models.dataset import DocumentSegment, Dataset
  9. @shared_task
  10. def clean_document_task(document_id: str, dataset_id: str):
  11. """
  12. Clean document when document deleted.
  13. :param document_id: document id
  14. :param dataset_id: dataset id
  15. Usage: clean_document_task.delay(document_id, dataset_id)
  16. """
  17. logging.info(click.style('Start clean document when document deleted: {}'.format(document_id), fg='green'))
  18. start_at = time.perf_counter()
  19. try:
  20. dataset = db.session.query(Dataset).filter(Dataset.id == dataset_id).first()
  21. if not dataset:
  22. raise Exception('Document has no dataset')
  23. vector_index = VectorIndex(dataset=dataset)
  24. keyword_table_index = KeywordTableIndex(dataset=dataset)
  25. segments = db.session.query(DocumentSegment).filter(DocumentSegment.document_id == document_id).all()
  26. index_node_ids = [segment.index_node_id for segment in segments]
  27. # delete from vector index
  28. if dataset.indexing_technique == "high_quality":
  29. vector_index.del_nodes(index_node_ids)
  30. # delete from keyword index
  31. if index_node_ids:
  32. keyword_table_index.del_nodes(index_node_ids)
  33. for segment in segments:
  34. db.session.delete(segment)
  35. end_at = time.perf_counter()
  36. logging.info(
  37. click.style('Cleaned document when document deleted: {} latency: {}'.format(document_id, end_at - start_at), fg='green'))
  38. except Exception:
  39. logging.exception("Cleaned document when document deleted failed")