|
@@ -27,32 +27,29 @@ from models.provider import Provider, ProviderModel
|
|
|
from services.account_service import RegisterService, TenantService
|
|
|
|
|
|
|
|
|
-@click.command('reset-password', help='Reset the account password.')
|
|
|
-@click.option('--email', prompt=True, help='The email address of the account whose password you need to reset')
|
|
|
-@click.option('--new-password', prompt=True, help='the new password.')
|
|
|
-@click.option('--password-confirm', prompt=True, help='the new password confirm.')
|
|
|
+@click.command("reset-password", help="Reset the account password.")
|
|
|
+@click.option("--email", prompt=True, help="The email address of the account whose password you need to reset")
|
|
|
+@click.option("--new-password", prompt=True, help="the new password.")
|
|
|
+@click.option("--password-confirm", prompt=True, help="the new password confirm.")
|
|
|
def reset_password(email, new_password, password_confirm):
|
|
|
"""
|
|
|
Reset password of owner account
|
|
|
Only available in SELF_HOSTED mode
|
|
|
"""
|
|
|
if str(new_password).strip() != str(password_confirm).strip():
|
|
|
- click.echo(click.style('sorry. The two passwords do not match.', fg='red'))
|
|
|
+ click.echo(click.style("sorry. The two passwords do not match.", fg="red"))
|
|
|
return
|
|
|
|
|
|
- account = db.session.query(Account). \
|
|
|
- filter(Account.email == email). \
|
|
|
- one_or_none()
|
|
|
+ account = db.session.query(Account).filter(Account.email == email).one_or_none()
|
|
|
|
|
|
if not account:
|
|
|
- click.echo(click.style('sorry. the account: [{}] not exist .'.format(email), fg='red'))
|
|
|
+ click.echo(click.style("sorry. the account: [{}] not exist .".format(email), fg="red"))
|
|
|
return
|
|
|
|
|
|
try:
|
|
|
valid_password(new_password)
|
|
|
except:
|
|
|
- click.echo(
|
|
|
- click.style('sorry. The passwords must match {} '.format(password_pattern), fg='red'))
|
|
|
+ click.echo(click.style("sorry. The passwords must match {} ".format(password_pattern), fg="red"))
|
|
|
return
|
|
|
|
|
|
# generate password salt
|
|
@@ -65,80 +62,87 @@ def reset_password(email, new_password, password_confirm):
|
|
|
account.password = base64_password_hashed
|
|
|
account.password_salt = base64_salt
|
|
|
db.session.commit()
|
|
|
- click.echo(click.style('Congratulations! Password has been reset.', fg='green'))
|
|
|
+ click.echo(click.style("Congratulations! Password has been reset.", fg="green"))
|
|
|
|
|
|
|
|
|
-@click.command('reset-email', help='Reset the account email.')
|
|
|
-@click.option('--email', prompt=True, help='The old email address of the account whose email you need to reset')
|
|
|
-@click.option('--new-email', prompt=True, help='the new email.')
|
|
|
-@click.option('--email-confirm', prompt=True, help='the new email confirm.')
|
|
|
+@click.command("reset-email", help="Reset the account email.")
|
|
|
+@click.option("--email", prompt=True, help="The old email address of the account whose email you need to reset")
|
|
|
+@click.option("--new-email", prompt=True, help="the new email.")
|
|
|
+@click.option("--email-confirm", prompt=True, help="the new email confirm.")
|
|
|
def reset_email(email, new_email, email_confirm):
|
|
|
"""
|
|
|
Replace account email
|
|
|
:return:
|
|
|
"""
|
|
|
if str(new_email).strip() != str(email_confirm).strip():
|
|
|
- click.echo(click.style('Sorry, new email and confirm email do not match.', fg='red'))
|
|
|
+ click.echo(click.style("Sorry, new email and confirm email do not match.", fg="red"))
|
|
|
return
|
|
|
|
|
|
- account = db.session.query(Account). \
|
|
|
- filter(Account.email == email). \
|
|
|
- one_or_none()
|
|
|
+ account = db.session.query(Account).filter(Account.email == email).one_or_none()
|
|
|
|
|
|
if not account:
|
|
|
- click.echo(click.style('sorry. the account: [{}] not exist .'.format(email), fg='red'))
|
|
|
+ click.echo(click.style("sorry. the account: [{}] not exist .".format(email), fg="red"))
|
|
|
return
|
|
|
|
|
|
try:
|
|
|
email_validate(new_email)
|
|
|
except:
|
|
|
- click.echo(
|
|
|
- click.style('sorry. {} is not a valid email. '.format(email), fg='red'))
|
|
|
+ click.echo(click.style("sorry. {} is not a valid email. ".format(email), fg="red"))
|
|
|
return
|
|
|
|
|
|
account.email = new_email
|
|
|
db.session.commit()
|
|
|
- click.echo(click.style('Congratulations!, email has been reset.', fg='green'))
|
|
|
-
|
|
|
-
|
|
|
-@click.command('reset-encrypt-key-pair', help='Reset the asymmetric key pair of workspace for encrypt LLM credentials. '
|
|
|
- 'After the reset, all LLM credentials will become invalid, '
|
|
|
- 'requiring re-entry.'
|
|
|
- 'Only support SELF_HOSTED mode.')
|
|
|
-@click.confirmation_option(prompt=click.style('Are you sure you want to reset encrypt key pair?'
|
|
|
- ' this operation cannot be rolled back!', fg='red'))
|
|
|
+ click.echo(click.style("Congratulations!, email has been reset.", fg="green"))
|
|
|
+
|
|
|
+
|
|
|
+@click.command(
|
|
|
+ "reset-encrypt-key-pair",
|
|
|
+ help="Reset the asymmetric key pair of workspace for encrypt LLM credentials. "
|
|
|
+ "After the reset, all LLM credentials will become invalid, "
|
|
|
+ "requiring re-entry."
|
|
|
+ "Only support SELF_HOSTED mode.",
|
|
|
+)
|
|
|
+@click.confirmation_option(
|
|
|
+ prompt=click.style(
|
|
|
+ "Are you sure you want to reset encrypt key pair?" " this operation cannot be rolled back!", fg="red"
|
|
|
+ )
|
|
|
+)
|
|
|
def reset_encrypt_key_pair():
|
|
|
"""
|
|
|
Reset the encrypted key pair of workspace for encrypt LLM credentials.
|
|
|
After the reset, all LLM credentials will become invalid, requiring re-entry.
|
|
|
Only support SELF_HOSTED mode.
|
|
|
"""
|
|
|
- if dify_config.EDITION != 'SELF_HOSTED':
|
|
|
- click.echo(click.style('Sorry, only support SELF_HOSTED mode.', fg='red'))
|
|
|
+ if dify_config.EDITION != "SELF_HOSTED":
|
|
|
+ click.echo(click.style("Sorry, only support SELF_HOSTED mode.", fg="red"))
|
|
|
return
|
|
|
|
|
|
tenants = db.session.query(Tenant).all()
|
|
|
for tenant in tenants:
|
|
|
if not tenant:
|
|
|
- click.echo(click.style('Sorry, no workspace found. Please enter /install to initialize.', fg='red'))
|
|
|
+ click.echo(click.style("Sorry, no workspace found. Please enter /install to initialize.", fg="red"))
|
|
|
return
|
|
|
|
|
|
tenant.encrypt_public_key = generate_key_pair(tenant.id)
|
|
|
|
|
|
- db.session.query(Provider).filter(Provider.provider_type == 'custom', Provider.tenant_id == tenant.id).delete()
|
|
|
+ db.session.query(Provider).filter(Provider.provider_type == "custom", Provider.tenant_id == tenant.id).delete()
|
|
|
db.session.query(ProviderModel).filter(ProviderModel.tenant_id == tenant.id).delete()
|
|
|
db.session.commit()
|
|
|
|
|
|
- click.echo(click.style('Congratulations! '
|
|
|
- 'the asymmetric key pair of workspace {} has been reset.'.format(tenant.id), fg='green'))
|
|
|
+ click.echo(
|
|
|
+ click.style(
|
|
|
+ "Congratulations! " "the asymmetric key pair of workspace {} has been reset.".format(tenant.id),
|
|
|
+ fg="green",
|
|
|
+ )
|
|
|
+ )
|
|
|
|
|
|
|
|
|
-@click.command('vdb-migrate', help='migrate vector db.')
|
|
|
-@click.option('--scope', default='all', prompt=False, help='The scope of vector database to migrate, Default is All.')
|
|
|
+@click.command("vdb-migrate", help="migrate vector db.")
|
|
|
+@click.option("--scope", default="all", prompt=False, help="The scope of vector database to migrate, Default is All.")
|
|
|
def vdb_migrate(scope: str):
|
|
|
- if scope in ['knowledge', 'all']:
|
|
|
+ if scope in ["knowledge", "all"]:
|
|
|
migrate_knowledge_vector_database()
|
|
|
- if scope in ['annotation', 'all']:
|
|
|
+ if scope in ["annotation", "all"]:
|
|
|
migrate_annotation_vector_database()
|
|
|
|
|
|
|
|
@@ -146,7 +150,7 @@ def migrate_annotation_vector_database():
|
|
|
"""
|
|
|
Migrate annotation datas to target vector database .
|
|
|
"""
|
|
|
- click.echo(click.style('Start migrate annotation data.', fg='green'))
|
|
|
+ click.echo(click.style("Start migrate annotation data.", fg="green"))
|
|
|
create_count = 0
|
|
|
skipped_count = 0
|
|
|
total_count = 0
|
|
@@ -154,98 +158,103 @@ def migrate_annotation_vector_database():
|
|
|
while True:
|
|
|
try:
|
|
|
# get apps info
|
|
|
- apps = db.session.query(App).filter(
|
|
|
- App.status == 'normal'
|
|
|
- ).order_by(App.created_at.desc()).paginate(page=page, per_page=50)
|
|
|
+ apps = (
|
|
|
+ db.session.query(App)
|
|
|
+ .filter(App.status == "normal")
|
|
|
+ .order_by(App.created_at.desc())
|
|
|
+ .paginate(page=page, per_page=50)
|
|
|
+ )
|
|
|
except NotFound:
|
|
|
break
|
|
|
|
|
|
page += 1
|
|
|
for app in apps:
|
|
|
total_count = total_count + 1
|
|
|
- click.echo(f'Processing the {total_count} app {app.id}. '
|
|
|
- + f'{create_count} created, {skipped_count} skipped.')
|
|
|
+ click.echo(
|
|
|
+ f"Processing the {total_count} app {app.id}. " + f"{create_count} created, {skipped_count} skipped."
|
|
|
+ )
|
|
|
try:
|
|
|
- click.echo('Create app annotation index: {}'.format(app.id))
|
|
|
- app_annotation_setting = db.session.query(AppAnnotationSetting).filter(
|
|
|
- AppAnnotationSetting.app_id == app.id
|
|
|
- ).first()
|
|
|
+ click.echo("Create app annotation index: {}".format(app.id))
|
|
|
+ app_annotation_setting = (
|
|
|
+ db.session.query(AppAnnotationSetting).filter(AppAnnotationSetting.app_id == app.id).first()
|
|
|
+ )
|
|
|
|
|
|
if not app_annotation_setting:
|
|
|
skipped_count = skipped_count + 1
|
|
|
- click.echo('App annotation setting is disabled: {}'.format(app.id))
|
|
|
+ click.echo("App annotation setting is disabled: {}".format(app.id))
|
|
|
continue
|
|
|
# get dataset_collection_binding info
|
|
|
- dataset_collection_binding = db.session.query(DatasetCollectionBinding).filter(
|
|
|
- DatasetCollectionBinding.id == app_annotation_setting.collection_binding_id
|
|
|
- ).first()
|
|
|
+ dataset_collection_binding = (
|
|
|
+ db.session.query(DatasetCollectionBinding)
|
|
|
+ .filter(DatasetCollectionBinding.id == app_annotation_setting.collection_binding_id)
|
|
|
+ .first()
|
|
|
+ )
|
|
|
if not dataset_collection_binding:
|
|
|
- click.echo('App annotation collection binding is not exist: {}'.format(app.id))
|
|
|
+ click.echo("App annotation collection binding is not exist: {}".format(app.id))
|
|
|
continue
|
|
|
annotations = db.session.query(MessageAnnotation).filter(MessageAnnotation.app_id == app.id).all()
|
|
|
dataset = Dataset(
|
|
|
id=app.id,
|
|
|
tenant_id=app.tenant_id,
|
|
|
- indexing_technique='high_quality',
|
|
|
+ indexing_technique="high_quality",
|
|
|
embedding_model_provider=dataset_collection_binding.provider_name,
|
|
|
embedding_model=dataset_collection_binding.model_name,
|
|
|
- collection_binding_id=dataset_collection_binding.id
|
|
|
+ collection_binding_id=dataset_collection_binding.id,
|
|
|
)
|
|
|
documents = []
|
|
|
if annotations:
|
|
|
for annotation in annotations:
|
|
|
document = Document(
|
|
|
page_content=annotation.question,
|
|
|
- metadata={
|
|
|
- "annotation_id": annotation.id,
|
|
|
- "app_id": app.id,
|
|
|
- "doc_id": annotation.id
|
|
|
- }
|
|
|
+ metadata={"annotation_id": annotation.id, "app_id": app.id, "doc_id": annotation.id},
|
|
|
)
|
|
|
documents.append(document)
|
|
|
|
|
|
- vector = Vector(dataset, attributes=['doc_id', 'annotation_id', 'app_id'])
|
|
|
+ vector = Vector(dataset, attributes=["doc_id", "annotation_id", "app_id"])
|
|
|
click.echo(f"Start to migrate annotation, app_id: {app.id}.")
|
|
|
|
|
|
try:
|
|
|
vector.delete()
|
|
|
- click.echo(
|
|
|
- click.style(f'Successfully delete vector index for app: {app.id}.',
|
|
|
- fg='green'))
|
|
|
+ click.echo(click.style(f"Successfully delete vector index for app: {app.id}.", fg="green"))
|
|
|
except Exception as e:
|
|
|
- click.echo(
|
|
|
- click.style(f'Failed to delete vector index for app {app.id}.',
|
|
|
- fg='red'))
|
|
|
+ click.echo(click.style(f"Failed to delete vector index for app {app.id}.", fg="red"))
|
|
|
raise e
|
|
|
if documents:
|
|
|
try:
|
|
|
- click.echo(click.style(
|
|
|
- f'Start to created vector index with {len(documents)} annotations for app {app.id}.',
|
|
|
- fg='green'))
|
|
|
- vector.create(documents)
|
|
|
click.echo(
|
|
|
- click.style(f'Successfully created vector index for app {app.id}.', fg='green'))
|
|
|
+ click.style(
|
|
|
+ f"Start to created vector index with {len(documents)} annotations for app {app.id}.",
|
|
|
+ fg="green",
|
|
|
+ )
|
|
|
+ )
|
|
|
+ vector.create(documents)
|
|
|
+ click.echo(click.style(f"Successfully created vector index for app {app.id}.", fg="green"))
|
|
|
except Exception as e:
|
|
|
- click.echo(click.style(f'Failed to created vector index for app {app.id}.', fg='red'))
|
|
|
+ click.echo(click.style(f"Failed to created vector index for app {app.id}.", fg="red"))
|
|
|
raise e
|
|
|
- click.echo(f'Successfully migrated app annotation {app.id}.')
|
|
|
+ click.echo(f"Successfully migrated app annotation {app.id}.")
|
|
|
create_count += 1
|
|
|
except Exception as e:
|
|
|
click.echo(
|
|
|
- click.style('Create app annotation index error: {} {}'.format(e.__class__.__name__, str(e)),
|
|
|
- fg='red'))
|
|
|
+ click.style(
|
|
|
+ "Create app annotation index error: {} {}".format(e.__class__.__name__, str(e)), fg="red"
|
|
|
+ )
|
|
|
+ )
|
|
|
continue
|
|
|
|
|
|
click.echo(
|
|
|
- click.style(f'Congratulations! Create {create_count} app annotation indexes, and skipped {skipped_count} apps.',
|
|
|
- fg='green'))
|
|
|
+ click.style(
|
|
|
+ f"Congratulations! Create {create_count} app annotation indexes, and skipped {skipped_count} apps.",
|
|
|
+ fg="green",
|
|
|
+ )
|
|
|
+ )
|
|
|
|
|
|
|
|
|
def migrate_knowledge_vector_database():
|
|
|
"""
|
|
|
Migrate vector database datas to target vector database .
|
|
|
"""
|
|
|
- click.echo(click.style('Start migrate vector db.', fg='green'))
|
|
|
+ click.echo(click.style("Start migrate vector db.", fg="green"))
|
|
|
create_count = 0
|
|
|
skipped_count = 0
|
|
|
total_count = 0
|
|
@@ -253,87 +262,77 @@ def migrate_knowledge_vector_database():
|
|
|
page = 1
|
|
|
while True:
|
|
|
try:
|
|
|
- datasets = db.session.query(Dataset).filter(Dataset.indexing_technique == 'high_quality') \
|
|
|
- .order_by(Dataset.created_at.desc()).paginate(page=page, per_page=50)
|
|
|
+ datasets = (
|
|
|
+ db.session.query(Dataset)
|
|
|
+ .filter(Dataset.indexing_technique == "high_quality")
|
|
|
+ .order_by(Dataset.created_at.desc())
|
|
|
+ .paginate(page=page, per_page=50)
|
|
|
+ )
|
|
|
except NotFound:
|
|
|
break
|
|
|
|
|
|
page += 1
|
|
|
for dataset in datasets:
|
|
|
total_count = total_count + 1
|
|
|
- click.echo(f'Processing the {total_count} dataset {dataset.id}. '
|
|
|
- + f'{create_count} created, {skipped_count} skipped.')
|
|
|
+ click.echo(
|
|
|
+ f"Processing the {total_count} dataset {dataset.id}. "
|
|
|
+ + f"{create_count} created, {skipped_count} skipped."
|
|
|
+ )
|
|
|
try:
|
|
|
- click.echo('Create dataset vdb index: {}'.format(dataset.id))
|
|
|
+ click.echo("Create dataset vdb index: {}".format(dataset.id))
|
|
|
if dataset.index_struct_dict:
|
|
|
- if dataset.index_struct_dict['type'] == vector_type:
|
|
|
+ if dataset.index_struct_dict["type"] == vector_type:
|
|
|
skipped_count = skipped_count + 1
|
|
|
continue
|
|
|
- collection_name = ''
|
|
|
+ collection_name = ""
|
|
|
if vector_type == VectorType.WEAVIATE:
|
|
|
dataset_id = dataset.id
|
|
|
collection_name = Dataset.gen_collection_name_by_id(dataset_id)
|
|
|
- index_struct_dict = {
|
|
|
- "type": VectorType.WEAVIATE,
|
|
|
- "vector_store": {"class_prefix": collection_name}
|
|
|
- }
|
|
|
+ index_struct_dict = {"type": VectorType.WEAVIATE, "vector_store": {"class_prefix": collection_name}}
|
|
|
dataset.index_struct = json.dumps(index_struct_dict)
|
|
|
elif vector_type == VectorType.QDRANT:
|
|
|
if dataset.collection_binding_id:
|
|
|
- dataset_collection_binding = db.session.query(DatasetCollectionBinding). \
|
|
|
- filter(DatasetCollectionBinding.id == dataset.collection_binding_id). \
|
|
|
- one_or_none()
|
|
|
+ dataset_collection_binding = (
|
|
|
+ db.session.query(DatasetCollectionBinding)
|
|
|
+ .filter(DatasetCollectionBinding.id == dataset.collection_binding_id)
|
|
|
+ .one_or_none()
|
|
|
+ )
|
|
|
if dataset_collection_binding:
|
|
|
collection_name = dataset_collection_binding.collection_name
|
|
|
else:
|
|
|
- raise ValueError('Dataset Collection Bindings is not exist!')
|
|
|
+ raise ValueError("Dataset Collection Bindings is not exist!")
|
|
|
else:
|
|
|
dataset_id = dataset.id
|
|
|
collection_name = Dataset.gen_collection_name_by_id(dataset_id)
|
|
|
- index_struct_dict = {
|
|
|
- "type": VectorType.QDRANT,
|
|
|
- "vector_store": {"class_prefix": collection_name}
|
|
|
- }
|
|
|
+ index_struct_dict = {"type": VectorType.QDRANT, "vector_store": {"class_prefix": collection_name}}
|
|
|
dataset.index_struct = json.dumps(index_struct_dict)
|
|
|
|
|
|
elif vector_type == VectorType.MILVUS:
|
|
|
dataset_id = dataset.id
|
|
|
collection_name = Dataset.gen_collection_name_by_id(dataset_id)
|
|
|
- index_struct_dict = {
|
|
|
- "type": VectorType.MILVUS,
|
|
|
- "vector_store": {"class_prefix": collection_name}
|
|
|
- }
|
|
|
+ index_struct_dict = {"type": VectorType.MILVUS, "vector_store": {"class_prefix": collection_name}}
|
|
|
dataset.index_struct = json.dumps(index_struct_dict)
|
|
|
elif vector_type == VectorType.RELYT:
|
|
|
dataset_id = dataset.id
|
|
|
collection_name = Dataset.gen_collection_name_by_id(dataset_id)
|
|
|
- index_struct_dict = {
|
|
|
- "type": 'relyt',
|
|
|
- "vector_store": {"class_prefix": collection_name}
|
|
|
- }
|
|
|
+ index_struct_dict = {"type": "relyt", "vector_store": {"class_prefix": collection_name}}
|
|
|
dataset.index_struct = json.dumps(index_struct_dict)
|
|
|
elif vector_type == VectorType.TENCENT:
|
|
|
dataset_id = dataset.id
|
|
|
collection_name = Dataset.gen_collection_name_by_id(dataset_id)
|
|
|
- index_struct_dict = {
|
|
|
- "type": VectorType.TENCENT,
|
|
|
- "vector_store": {"class_prefix": collection_name}
|
|
|
- }
|
|
|
+ index_struct_dict = {"type": VectorType.TENCENT, "vector_store": {"class_prefix": collection_name}}
|
|
|
dataset.index_struct = json.dumps(index_struct_dict)
|
|
|
elif vector_type == VectorType.PGVECTOR:
|
|
|
dataset_id = dataset.id
|
|
|
collection_name = Dataset.gen_collection_name_by_id(dataset_id)
|
|
|
- index_struct_dict = {
|
|
|
- "type": VectorType.PGVECTOR,
|
|
|
- "vector_store": {"class_prefix": collection_name}
|
|
|
- }
|
|
|
+ index_struct_dict = {"type": VectorType.PGVECTOR, "vector_store": {"class_prefix": collection_name}}
|
|
|
dataset.index_struct = json.dumps(index_struct_dict)
|
|
|
elif vector_type == VectorType.OPENSEARCH:
|
|
|
dataset_id = dataset.id
|
|
|
collection_name = Dataset.gen_collection_name_by_id(dataset_id)
|
|
|
index_struct_dict = {
|
|
|
"type": VectorType.OPENSEARCH,
|
|
|
- "vector_store": {"class_prefix": collection_name}
|
|
|
+ "vector_store": {"class_prefix": collection_name},
|
|
|
}
|
|
|
dataset.index_struct = json.dumps(index_struct_dict)
|
|
|
elif vector_type == VectorType.ANALYTICDB:
|
|
@@ -341,16 +340,13 @@ def migrate_knowledge_vector_database():
|
|
|
collection_name = Dataset.gen_collection_name_by_id(dataset_id)
|
|
|
index_struct_dict = {
|
|
|
"type": VectorType.ANALYTICDB,
|
|
|
- "vector_store": {"class_prefix": collection_name}
|
|
|
+ "vector_store": {"class_prefix": collection_name},
|
|
|
}
|
|
|
dataset.index_struct = json.dumps(index_struct_dict)
|
|
|
elif vector_type == VectorType.ELASTICSEARCH:
|
|
|
dataset_id = dataset.id
|
|
|
index_name = Dataset.gen_collection_name_by_id(dataset_id)
|
|
|
- index_struct_dict = {
|
|
|
- "type": 'elasticsearch',
|
|
|
- "vector_store": {"class_prefix": index_name}
|
|
|
- }
|
|
|
+ index_struct_dict = {"type": "elasticsearch", "vector_store": {"class_prefix": index_name}}
|
|
|
dataset.index_struct = json.dumps(index_struct_dict)
|
|
|
else:
|
|
|
raise ValueError(f"Vector store {vector_type} is not supported.")
|
|
@@ -361,29 +357,41 @@ def migrate_knowledge_vector_database():
|
|
|
try:
|
|
|
vector.delete()
|
|
|
click.echo(
|
|
|
- click.style(f'Successfully delete vector index {collection_name} for dataset {dataset.id}.',
|
|
|
- fg='green'))
|
|
|
+ click.style(
|
|
|
+ f"Successfully delete vector index {collection_name} for dataset {dataset.id}.", fg="green"
|
|
|
+ )
|
|
|
+ )
|
|
|
except Exception as e:
|
|
|
click.echo(
|
|
|
- click.style(f'Failed to delete vector index {collection_name} for dataset {dataset.id}.',
|
|
|
- fg='red'))
|
|
|
+ click.style(
|
|
|
+ f"Failed to delete vector index {collection_name} for dataset {dataset.id}.", fg="red"
|
|
|
+ )
|
|
|
+ )
|
|
|
raise e
|
|
|
|
|
|
- dataset_documents = db.session.query(DatasetDocument).filter(
|
|
|
- DatasetDocument.dataset_id == dataset.id,
|
|
|
- DatasetDocument.indexing_status == 'completed',
|
|
|
- DatasetDocument.enabled == True,
|
|
|
- DatasetDocument.archived == False,
|
|
|
- ).all()
|
|
|
+ dataset_documents = (
|
|
|
+ db.session.query(DatasetDocument)
|
|
|
+ .filter(
|
|
|
+ DatasetDocument.dataset_id == dataset.id,
|
|
|
+ DatasetDocument.indexing_status == "completed",
|
|
|
+ DatasetDocument.enabled == True,
|
|
|
+ DatasetDocument.archived == False,
|
|
|
+ )
|
|
|
+ .all()
|
|
|
+ )
|
|
|
|
|
|
documents = []
|
|
|
segments_count = 0
|
|
|
for dataset_document in dataset_documents:
|
|
|
- segments = db.session.query(DocumentSegment).filter(
|
|
|
- DocumentSegment.document_id == dataset_document.id,
|
|
|
- DocumentSegment.status == 'completed',
|
|
|
- DocumentSegment.enabled == True
|
|
|
- ).all()
|
|
|
+ segments = (
|
|
|
+ db.session.query(DocumentSegment)
|
|
|
+ .filter(
|
|
|
+ DocumentSegment.document_id == dataset_document.id,
|
|
|
+ DocumentSegment.status == "completed",
|
|
|
+ DocumentSegment.enabled == True,
|
|
|
+ )
|
|
|
+ .all()
|
|
|
+ )
|
|
|
|
|
|
for segment in segments:
|
|
|
document = Document(
|
|
@@ -393,7 +401,7 @@ def migrate_knowledge_vector_database():
|
|
|
"doc_hash": segment.index_node_hash,
|
|
|
"document_id": segment.document_id,
|
|
|
"dataset_id": segment.dataset_id,
|
|
|
- }
|
|
|
+ },
|
|
|
)
|
|
|
|
|
|
documents.append(document)
|
|
@@ -401,37 +409,43 @@ def migrate_knowledge_vector_database():
|
|
|
|
|
|
if documents:
|
|
|
try:
|
|
|
- click.echo(click.style(
|
|
|
- f'Start to created vector index with {len(documents)} documents of {segments_count} segments for dataset {dataset.id}.',
|
|
|
- fg='green'))
|
|
|
+ click.echo(
|
|
|
+ click.style(
|
|
|
+ f"Start to created vector index with {len(documents)} documents of {segments_count} segments for dataset {dataset.id}.",
|
|
|
+ fg="green",
|
|
|
+ )
|
|
|
+ )
|
|
|
vector.create(documents)
|
|
|
click.echo(
|
|
|
- click.style(f'Successfully created vector index for dataset {dataset.id}.', fg='green'))
|
|
|
+ click.style(f"Successfully created vector index for dataset {dataset.id}.", fg="green")
|
|
|
+ )
|
|
|
except Exception as e:
|
|
|
- click.echo(click.style(f'Failed to created vector index for dataset {dataset.id}.', fg='red'))
|
|
|
+ click.echo(click.style(f"Failed to created vector index for dataset {dataset.id}.", fg="red"))
|
|
|
raise e
|
|
|
db.session.add(dataset)
|
|
|
db.session.commit()
|
|
|
- click.echo(f'Successfully migrated dataset {dataset.id}.')
|
|
|
+ click.echo(f"Successfully migrated dataset {dataset.id}.")
|
|
|
create_count += 1
|
|
|
except Exception as e:
|
|
|
db.session.rollback()
|
|
|
click.echo(
|
|
|
- click.style('Create dataset index error: {} {}'.format(e.__class__.__name__, str(e)),
|
|
|
- fg='red'))
|
|
|
+ click.style("Create dataset index error: {} {}".format(e.__class__.__name__, str(e)), fg="red")
|
|
|
+ )
|
|
|
continue
|
|
|
|
|
|
click.echo(
|
|
|
- click.style(f'Congratulations! Create {create_count} dataset indexes, and skipped {skipped_count} datasets.',
|
|
|
- fg='green'))
|
|
|
+ click.style(
|
|
|
+ f"Congratulations! Create {create_count} dataset indexes, and skipped {skipped_count} datasets.", fg="green"
|
|
|
+ )
|
|
|
+ )
|
|
|
|
|
|
|
|
|
-@click.command('convert-to-agent-apps', help='Convert Agent Assistant to Agent App.')
|
|
|
+@click.command("convert-to-agent-apps", help="Convert Agent Assistant to Agent App.")
|
|
|
def convert_to_agent_apps():
|
|
|
"""
|
|
|
Convert Agent Assistant to Agent App.
|
|
|
"""
|
|
|
- click.echo(click.style('Start convert to agent apps.', fg='green'))
|
|
|
+ click.echo(click.style("Start convert to agent apps.", fg="green"))
|
|
|
|
|
|
proceeded_app_ids = []
|
|
|
|
|
@@ -466,7 +480,7 @@ def convert_to_agent_apps():
|
|
|
break
|
|
|
|
|
|
for app in apps:
|
|
|
- click.echo('Converting app: {}'.format(app.id))
|
|
|
+ click.echo("Converting app: {}".format(app.id))
|
|
|
|
|
|
try:
|
|
|
app.mode = AppMode.AGENT_CHAT.value
|
|
@@ -478,137 +492,139 @@ def convert_to_agent_apps():
|
|
|
)
|
|
|
|
|
|
db.session.commit()
|
|
|
- click.echo(click.style('Converted app: {}'.format(app.id), fg='green'))
|
|
|
+ click.echo(click.style("Converted app: {}".format(app.id), fg="green"))
|
|
|
except Exception as e:
|
|
|
- click.echo(
|
|
|
- click.style('Convert app error: {} {}'.format(e.__class__.__name__,
|
|
|
- str(e)), fg='red'))
|
|
|
+ click.echo(click.style("Convert app error: {} {}".format(e.__class__.__name__, str(e)), fg="red"))
|
|
|
|
|
|
- click.echo(click.style('Congratulations! Converted {} agent apps.'.format(len(proceeded_app_ids)), fg='green'))
|
|
|
+ click.echo(click.style("Congratulations! Converted {} agent apps.".format(len(proceeded_app_ids)), fg="green"))
|
|
|
|
|
|
|
|
|
-@click.command('add-qdrant-doc-id-index', help='add qdrant doc_id index.')
|
|
|
-@click.option('--field', default='metadata.doc_id', prompt=False, help='index field , default is metadata.doc_id.')
|
|
|
+@click.command("add-qdrant-doc-id-index", help="add qdrant doc_id index.")
|
|
|
+@click.option("--field", default="metadata.doc_id", prompt=False, help="index field , default is metadata.doc_id.")
|
|
|
def add_qdrant_doc_id_index(field: str):
|
|
|
- click.echo(click.style('Start add qdrant doc_id index.', fg='green'))
|
|
|
+ click.echo(click.style("Start add qdrant doc_id index.", fg="green"))
|
|
|
vector_type = dify_config.VECTOR_STORE
|
|
|
if vector_type != "qdrant":
|
|
|
- click.echo(click.style('Sorry, only support qdrant vector store.', fg='red'))
|
|
|
+ click.echo(click.style("Sorry, only support qdrant vector store.", fg="red"))
|
|
|
return
|
|
|
create_count = 0
|
|
|
|
|
|
try:
|
|
|
bindings = db.session.query(DatasetCollectionBinding).all()
|
|
|
if not bindings:
|
|
|
- click.echo(click.style('Sorry, no dataset collection bindings found.', fg='red'))
|
|
|
+ click.echo(click.style("Sorry, no dataset collection bindings found.", fg="red"))
|
|
|
return
|
|
|
import qdrant_client
|
|
|
from qdrant_client.http.exceptions import UnexpectedResponse
|
|
|
from qdrant_client.http.models import PayloadSchemaType
|
|
|
|
|
|
from core.rag.datasource.vdb.qdrant.qdrant_vector import QdrantConfig
|
|
|
+
|
|
|
for binding in bindings:
|
|
|
if dify_config.QDRANT_URL is None:
|
|
|
- raise ValueError('Qdrant url is required.')
|
|
|
+ raise ValueError("Qdrant url is required.")
|
|
|
qdrant_config = QdrantConfig(
|
|
|
endpoint=dify_config.QDRANT_URL,
|
|
|
api_key=dify_config.QDRANT_API_KEY,
|
|
|
root_path=current_app.root_path,
|
|
|
timeout=dify_config.QDRANT_CLIENT_TIMEOUT,
|
|
|
grpc_port=dify_config.QDRANT_GRPC_PORT,
|
|
|
- prefer_grpc=dify_config.QDRANT_GRPC_ENABLED
|
|
|
+ prefer_grpc=dify_config.QDRANT_GRPC_ENABLED,
|
|
|
)
|
|
|
try:
|
|
|
client = qdrant_client.QdrantClient(**qdrant_config.to_qdrant_params())
|
|
|
# create payload index
|
|
|
- client.create_payload_index(binding.collection_name, field,
|
|
|
- field_schema=PayloadSchemaType.KEYWORD)
|
|
|
+ client.create_payload_index(binding.collection_name, field, field_schema=PayloadSchemaType.KEYWORD)
|
|
|
create_count += 1
|
|
|
except UnexpectedResponse as e:
|
|
|
# Collection does not exist, so return
|
|
|
if e.status_code == 404:
|
|
|
- click.echo(click.style(f'Collection not found, collection_name:{binding.collection_name}.', fg='red'))
|
|
|
+ click.echo(
|
|
|
+ click.style(f"Collection not found, collection_name:{binding.collection_name}.", fg="red")
|
|
|
+ )
|
|
|
continue
|
|
|
# Some other error occurred, so re-raise the exception
|
|
|
else:
|
|
|
- click.echo(click.style(f'Failed to create qdrant index, collection_name:{binding.collection_name}.', fg='red'))
|
|
|
+ click.echo(
|
|
|
+ click.style(
|
|
|
+ f"Failed to create qdrant index, collection_name:{binding.collection_name}.", fg="red"
|
|
|
+ )
|
|
|
+ )
|
|
|
|
|
|
except Exception as e:
|
|
|
- click.echo(click.style('Failed to create qdrant client.', fg='red'))
|
|
|
+ click.echo(click.style("Failed to create qdrant client.", fg="red"))
|
|
|
|
|
|
- click.echo(
|
|
|
- click.style(f'Congratulations! Create {create_count} collection indexes.',
|
|
|
- fg='green'))
|
|
|
+ click.echo(click.style(f"Congratulations! Create {create_count} collection indexes.", fg="green"))
|
|
|
|
|
|
|
|
|
-@click.command('create-tenant', help='Create account and tenant.')
|
|
|
-@click.option('--email', prompt=True, help='The email address of the tenant account.')
|
|
|
-@click.option('--language', prompt=True, help='Account language, default: en-US.')
|
|
|
+@click.command("create-tenant", help="Create account and tenant.")
|
|
|
+@click.option("--email", prompt=True, help="The email address of the tenant account.")
|
|
|
+@click.option("--language", prompt=True, help="Account language, default: en-US.")
|
|
|
def create_tenant(email: str, language: Optional[str] = None):
|
|
|
"""
|
|
|
Create tenant account
|
|
|
"""
|
|
|
if not email:
|
|
|
- click.echo(click.style('Sorry, email is required.', fg='red'))
|
|
|
+ click.echo(click.style("Sorry, email is required.", fg="red"))
|
|
|
return
|
|
|
|
|
|
# Create account
|
|
|
email = email.strip()
|
|
|
|
|
|
- if '@' not in email:
|
|
|
- click.echo(click.style('Sorry, invalid email address.', fg='red'))
|
|
|
+ if "@" not in email:
|
|
|
+ click.echo(click.style("Sorry, invalid email address.", fg="red"))
|
|
|
return
|
|
|
|
|
|
- account_name = email.split('@')[0]
|
|
|
+ account_name = email.split("@")[0]
|
|
|
|
|
|
if language not in languages:
|
|
|
- language = 'en-US'
|
|
|
+ language = "en-US"
|
|
|
|
|
|
# generate random password
|
|
|
new_password = secrets.token_urlsafe(16)
|
|
|
|
|
|
# register account
|
|
|
- account = RegisterService.register(
|
|
|
- email=email,
|
|
|
- name=account_name,
|
|
|
- password=new_password,
|
|
|
- language=language
|
|
|
- )
|
|
|
+ account = RegisterService.register(email=email, name=account_name, password=new_password, language=language)
|
|
|
|
|
|
TenantService.create_owner_tenant_if_not_exist(account)
|
|
|
|
|
|
- click.echo(click.style('Congratulations! Account and tenant created.\n'
|
|
|
- 'Account: {}\nPassword: {}'.format(email, new_password), fg='green'))
|
|
|
+ click.echo(
|
|
|
+ click.style(
|
|
|
+ "Congratulations! Account and tenant created.\n" "Account: {}\nPassword: {}".format(email, new_password),
|
|
|
+ fg="green",
|
|
|
+ )
|
|
|
+ )
|
|
|
|
|
|
|
|
|
-@click.command('upgrade-db', help='upgrade the database')
|
|
|
+@click.command("upgrade-db", help="upgrade the database")
|
|
|
def upgrade_db():
|
|
|
- click.echo('Preparing database migration...')
|
|
|
- lock = redis_client.lock(name='db_upgrade_lock', timeout=60)
|
|
|
+ click.echo("Preparing database migration...")
|
|
|
+ lock = redis_client.lock(name="db_upgrade_lock", timeout=60)
|
|
|
if lock.acquire(blocking=False):
|
|
|
try:
|
|
|
- click.echo(click.style('Start database migration.', fg='green'))
|
|
|
+ click.echo(click.style("Start database migration.", fg="green"))
|
|
|
|
|
|
# run db migration
|
|
|
import flask_migrate
|
|
|
+
|
|
|
flask_migrate.upgrade()
|
|
|
|
|
|
- click.echo(click.style('Database migration successful!', fg='green'))
|
|
|
+ click.echo(click.style("Database migration successful!", fg="green"))
|
|
|
|
|
|
except Exception as e:
|
|
|
- logging.exception(f'Database migration failed, error: {e}')
|
|
|
+ logging.exception(f"Database migration failed, error: {e}")
|
|
|
finally:
|
|
|
lock.release()
|
|
|
else:
|
|
|
- click.echo('Database migration skipped')
|
|
|
+ click.echo("Database migration skipped")
|
|
|
|
|
|
|
|
|
-@click.command('fix-app-site-missing', help='Fix app related site missing issue.')
|
|
|
+@click.command("fix-app-site-missing", help="Fix app related site missing issue.")
|
|
|
def fix_app_site_missing():
|
|
|
"""
|
|
|
Fix app related site missing issue.
|
|
|
"""
|
|
|
- click.echo(click.style('Start fix app related site missing issue.', fg='green'))
|
|
|
+ click.echo(click.style("Start fix app related site missing issue.", fg="green"))
|
|
|
|
|
|
failed_app_ids = []
|
|
|
while True:
|
|
@@ -639,15 +655,14 @@ where sites.id is null limit 1000"""
|
|
|
app_was_created.send(app, account=account)
|
|
|
except Exception as e:
|
|
|
failed_app_ids.append(app_id)
|
|
|
- click.echo(click.style('Fix app {} related site missing issue failed!'.format(app_id), fg='red'))
|
|
|
- logging.exception(f'Fix app related site missing issue failed, error: {e}')
|
|
|
+ click.echo(click.style("Fix app {} related site missing issue failed!".format(app_id), fg="red"))
|
|
|
+ logging.exception(f"Fix app related site missing issue failed, error: {e}")
|
|
|
continue
|
|
|
|
|
|
if not processed_count:
|
|
|
break
|
|
|
|
|
|
-
|
|
|
- click.echo(click.style('Congratulations! Fix app related site missing issue successful!', fg='green'))
|
|
|
+ click.echo(click.style("Congratulations! Fix app related site missing issue successful!", fg="green"))
|
|
|
|
|
|
|
|
|
def register_commands(app):
|