import base64 import json import secrets import click from flask import current_app from werkzeug.exceptions import NotFound from core.rag.datasource.vdb.vector_factory import Vector from core.rag.models.document import Document from extensions.ext_database import db from libs.helper import email as email_validate from libs.password import hash_password, password_pattern, valid_password from libs.rsa import generate_key_pair from models.account import Tenant from models.dataset import Dataset, DatasetCollectionBinding, DocumentSegment from models.dataset import Document as DatasetDocument from models.model import Account, App, AppAnnotationSetting, AppMode, Conversation, MessageAnnotation from models.provider import Provider, ProviderModel @click.command('reset-password', help='Reset the account password.') @click.option('--email', prompt=True, help='The email address of the account whose password you need to reset') @click.option('--new-password', prompt=True, help='the new password.') @click.option('--password-confirm', prompt=True, help='the new password confirm.') def reset_password(email, new_password, password_confirm): """ Reset password of owner account Only available in SELF_HOSTED mode """ if str(new_password).strip() != str(password_confirm).strip(): click.echo(click.style('sorry. The two passwords do not match.', fg='red')) return account = db.session.query(Account). \ filter(Account.email == email). \ one_or_none() if not account: click.echo(click.style('sorry. the account: [{}] not exist .'.format(email), fg='red')) return try: valid_password(new_password) except: click.echo( click.style('sorry. The passwords must match {} '.format(password_pattern), fg='red')) return # generate password salt salt = secrets.token_bytes(16) base64_salt = base64.b64encode(salt).decode() # encrypt password with salt password_hashed = hash_password(new_password, salt) base64_password_hashed = base64.b64encode(password_hashed).decode() account.password = base64_password_hashed account.password_salt = base64_salt db.session.commit() click.echo(click.style('Congratulations!, password has been reset.', fg='green')) @click.command('reset-email', help='Reset the account email.') @click.option('--email', prompt=True, help='The old email address of the account whose email you need to reset') @click.option('--new-email', prompt=True, help='the new email.') @click.option('--email-confirm', prompt=True, help='the new email confirm.') def reset_email(email, new_email, email_confirm): """ Replace account email :return: """ if str(new_email).strip() != str(email_confirm).strip(): click.echo(click.style('Sorry, new email and confirm email do not match.', fg='red')) return account = db.session.query(Account). \ filter(Account.email == email). \ one_or_none() if not account: click.echo(click.style('sorry. the account: [{}] not exist .'.format(email), fg='red')) return try: email_validate(new_email) except: click.echo( click.style('sorry. {} is not a valid email. '.format(email), fg='red')) return account.email = new_email db.session.commit() click.echo(click.style('Congratulations!, email has been reset.', fg='green')) @click.command('reset-encrypt-key-pair', help='Reset the asymmetric key pair of workspace for encrypt LLM credentials. ' 'After the reset, all LLM credentials will become invalid, ' 'requiring re-entry.' 'Only support SELF_HOSTED mode.') @click.confirmation_option(prompt=click.style('Are you sure you want to reset encrypt key pair?' ' this operation cannot be rolled back!', fg='red')) def reset_encrypt_key_pair(): """ Reset the encrypted key pair of workspace for encrypt LLM credentials. After the reset, all LLM credentials will become invalid, requiring re-entry. Only support SELF_HOSTED mode. """ if current_app.config['EDITION'] != 'SELF_HOSTED': click.echo(click.style('Sorry, only support SELF_HOSTED mode.', fg='red')) return tenants = db.session.query(Tenant).all() for tenant in tenants: if not tenant: click.echo(click.style('Sorry, no workspace found. Please enter /install to initialize.', fg='red')) return tenant.encrypt_public_key = generate_key_pair(tenant.id) db.session.query(Provider).filter(Provider.provider_type == 'custom', Provider.tenant_id == tenant.id).delete() db.session.query(ProviderModel).filter(ProviderModel.tenant_id == tenant.id).delete() db.session.commit() click.echo(click.style('Congratulations! ' 'the asymmetric key pair of workspace {} has been reset.'.format(tenant.id), fg='green')) @click.command('vdb-migrate', help='migrate vector db.') @click.option('--scope', default='all', prompt=False, help='The scope of vector database to migrate, Default is All.') def vdb_migrate(scope: str): if scope in ['knowledge', 'all']: migrate_knowledge_vector_database() if scope in ['annotation', 'all']: migrate_annotation_vector_database() def migrate_annotation_vector_database(): """ Migrate annotation datas to target vector database . """ click.echo(click.style('Start migrate annotation data.', fg='green')) create_count = 0 skipped_count = 0 total_count = 0 page = 1 while True: try: # get apps info apps = db.session.query(App).filter( App.status == 'normal' ).order_by(App.created_at.desc()).paginate(page=page, per_page=50) except NotFound: break page += 1 for app in apps: total_count = total_count + 1 click.echo(f'Processing the {total_count} app {app.id}. ' + f'{create_count} created, {skipped_count} skipped.') try: click.echo('Create app annotation index: {}'.format(app.id)) app_annotation_setting = db.session.query(AppAnnotationSetting).filter( AppAnnotationSetting.app_id == app.id ).first() if not app_annotation_setting: skipped_count = skipped_count + 1 click.echo('App annotation setting is disabled: {}'.format(app.id)) continue # get dataset_collection_binding info dataset_collection_binding = db.session.query(DatasetCollectionBinding).filter( DatasetCollectionBinding.id == app_annotation_setting.collection_binding_id ).first() if not dataset_collection_binding: click.echo('App annotation collection binding is not exist: {}'.format(app.id)) continue annotations = db.session.query(MessageAnnotation).filter(MessageAnnotation.app_id == app.id).all() dataset = Dataset( id=app.id, tenant_id=app.tenant_id, indexing_technique='high_quality', embedding_model_provider=dataset_collection_binding.provider_name, embedding_model=dataset_collection_binding.model_name, collection_binding_id=dataset_collection_binding.id ) documents = [] if annotations: for annotation in annotations: document = Document( page_content=annotation.question, metadata={ "annotation_id": annotation.id, "app_id": app.id, "doc_id": annotation.id } ) documents.append(document) vector = Vector(dataset, attributes=['doc_id', 'annotation_id', 'app_id']) click.echo(f"Start to migrate annotation, app_id: {app.id}.") try: vector.delete() click.echo( click.style(f'Successfully delete vector index for app: {app.id}.', fg='green')) except Exception as e: click.echo( click.style(f'Failed to delete vector index for app {app.id}.', fg='red')) raise e if documents: try: click.echo(click.style( f'Start to created vector index with {len(documents)} annotations for app {app.id}.', fg='green')) vector.create(documents) click.echo( click.style(f'Successfully created vector index for app {app.id}.', fg='green')) except Exception as e: click.echo(click.style(f'Failed to created vector index for app {app.id}.', fg='red')) raise e click.echo(f'Successfully migrated app annotation {app.id}.') create_count += 1 except Exception as e: click.echo( click.style('Create app annotation index error: {} {}'.format(e.__class__.__name__, str(e)), fg='red')) continue click.echo( click.style(f'Congratulations! Create {create_count} app annotation indexes, and skipped {skipped_count} apps.', fg='green')) def migrate_knowledge_vector_database(): """ Migrate vector database datas to target vector database . """ click.echo(click.style('Start migrate vector db.', fg='green')) create_count = 0 skipped_count = 0 total_count = 0 config = current_app.config vector_type = config.get('VECTOR_STORE') page = 1 while True: try: datasets = db.session.query(Dataset).filter(Dataset.indexing_technique == 'high_quality') \ .order_by(Dataset.created_at.desc()).paginate(page=page, per_page=50) except NotFound: break page += 1 for dataset in datasets: total_count = total_count + 1 click.echo(f'Processing the {total_count} dataset {dataset.id}. ' + f'{create_count} created, {skipped_count} skipped.') try: click.echo('Create dataset vdb index: {}'.format(dataset.id)) if dataset.index_struct_dict: if dataset.index_struct_dict['type'] == vector_type: skipped_count = skipped_count + 1 continue collection_name = '' if vector_type == "weaviate": dataset_id = dataset.id collection_name = Dataset.gen_collection_name_by_id(dataset_id) index_struct_dict = { "type": 'weaviate', "vector_store": {"class_prefix": collection_name} } dataset.index_struct = json.dumps(index_struct_dict) elif vector_type == "qdrant": if dataset.collection_binding_id: dataset_collection_binding = db.session.query(DatasetCollectionBinding). \ filter(DatasetCollectionBinding.id == dataset.collection_binding_id). \ one_or_none() if dataset_collection_binding: collection_name = dataset_collection_binding.collection_name else: raise ValueError('Dataset Collection Bindings is not exist!') else: dataset_id = dataset.id collection_name = Dataset.gen_collection_name_by_id(dataset_id) index_struct_dict = { "type": 'qdrant', "vector_store": {"class_prefix": collection_name} } dataset.index_struct = json.dumps(index_struct_dict) elif vector_type == "milvus": dataset_id = dataset.id collection_name = Dataset.gen_collection_name_by_id(dataset_id) index_struct_dict = { "type": 'milvus', "vector_store": {"class_prefix": collection_name} } dataset.index_struct = json.dumps(index_struct_dict) elif vector_type == "relyt": dataset_id = dataset.id collection_name = Dataset.gen_collection_name_by_id(dataset_id) index_struct_dict = { "type": 'relyt', "vector_store": {"class_prefix": collection_name} } dataset.index_struct = json.dumps(index_struct_dict) elif vector_type == "pgvector": dataset_id = dataset.id collection_name = Dataset.gen_collection_name_by_id(dataset_id) index_struct_dict = { "type": 'pgvector', "vector_store": {"class_prefix": collection_name} } dataset.index_struct = json.dumps(index_struct_dict) else: raise ValueError(f"Vector store {config.get('VECTOR_STORE')} is not supported.") vector = Vector(dataset) click.echo(f"Start to migrate dataset {dataset.id}.") try: vector.delete() click.echo( click.style(f'Successfully delete vector index {collection_name} for dataset {dataset.id}.', fg='green')) except Exception as e: click.echo( click.style(f'Failed to delete vector index {collection_name} for dataset {dataset.id}.', fg='red')) raise e dataset_documents = db.session.query(DatasetDocument).filter( DatasetDocument.dataset_id == dataset.id, DatasetDocument.indexing_status == 'completed', DatasetDocument.enabled == True, DatasetDocument.archived == False, ).all() documents = [] segments_count = 0 for dataset_document in dataset_documents: segments = db.session.query(DocumentSegment).filter( DocumentSegment.document_id == dataset_document.id, DocumentSegment.status == 'completed', DocumentSegment.enabled == True ).all() for segment in segments: document = Document( page_content=segment.content, metadata={ "doc_id": segment.index_node_id, "doc_hash": segment.index_node_hash, "document_id": segment.document_id, "dataset_id": segment.dataset_id, } ) documents.append(document) segments_count = segments_count + 1 if documents: try: click.echo(click.style( f'Start to created vector index with {len(documents)} documents of {segments_count} segments for dataset {dataset.id}.', fg='green')) vector.create(documents) click.echo( click.style(f'Successfully created vector index for dataset {dataset.id}.', fg='green')) except Exception as e: click.echo(click.style(f'Failed to created vector index for dataset {dataset.id}.', fg='red')) raise e db.session.add(dataset) db.session.commit() click.echo(f'Successfully migrated dataset {dataset.id}.') create_count += 1 except Exception as e: db.session.rollback() click.echo( click.style('Create dataset index error: {} {}'.format(e.__class__.__name__, str(e)), fg='red')) continue click.echo( click.style(f'Congratulations! Create {create_count} dataset indexes, and skipped {skipped_count} datasets.', fg='green')) @click.command('convert-to-agent-apps', help='Convert Agent Assistant to Agent App.') def convert_to_agent_apps(): """ Convert Agent Assistant to Agent App. """ click.echo(click.style('Start convert to agent apps.', fg='green')) proceeded_app_ids = [] while True: # fetch first 1000 apps sql_query = """SELECT a.id AS id FROM apps a INNER JOIN app_model_configs am ON a.app_model_config_id=am.id WHERE a.mode = 'chat' AND am.agent_mode is not null AND ( am.agent_mode like '%"strategy": "function_call"%' OR am.agent_mode like '%"strategy": "react"%' ) AND ( am.agent_mode like '{"enabled": true%' OR am.agent_mode like '{"max_iteration": %' ) ORDER BY a.created_at DESC LIMIT 1000 """ with db.engine.begin() as conn: rs = conn.execute(db.text(sql_query)) apps = [] for i in rs: app_id = str(i.id) if app_id not in proceeded_app_ids: proceeded_app_ids.append(app_id) app = db.session.query(App).filter(App.id == app_id).first() apps.append(app) if len(apps) == 0: break for app in apps: click.echo('Converting app: {}'.format(app.id)) try: app.mode = AppMode.AGENT_CHAT.value db.session.commit() # update conversation mode to agent db.session.query(Conversation).filter(Conversation.app_id == app.id).update( {Conversation.mode: AppMode.AGENT_CHAT.value} ) db.session.commit() click.echo(click.style('Converted app: {}'.format(app.id), fg='green')) except Exception as e: click.echo( click.style('Convert app error: {} {}'.format(e.__class__.__name__, str(e)), fg='red')) click.echo(click.style('Congratulations! Converted {} agent apps.'.format(len(proceeded_app_ids)), fg='green')) def register_commands(app): app.cli.add_command(reset_password) app.cli.add_command(reset_email) app.cli.add_command(reset_encrypt_key_pair) app.cli.add_command(vdb_migrate) app.cli.add_command(convert_to_agent_apps)