|  | @@ -0,0 +1,136 @@
 | 
	
		
			
				|  |  | +import json
 | 
	
		
			
				|  |  | +import os
 | 
	
		
			
				|  |  | +from unittest.mock import patch, MagicMock
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +from core.model_providers.models.entity.model_params import ModelType
 | 
	
		
			
				|  |  | +from core.model_providers.models.embedding.huggingface_embedding import HuggingfaceEmbedding
 | 
	
		
			
				|  |  | +from core.model_providers.providers.huggingface_hub_provider import HuggingfaceHubProvider
 | 
	
		
			
				|  |  | +from models.provider import Provider, ProviderType, ProviderModel
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +DEFAULT_MODEL_NAME = 'obrizum/all-MiniLM-L6-v2'
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +def get_mock_provider():
 | 
	
		
			
				|  |  | +    return Provider(
 | 
	
		
			
				|  |  | +        id='provider_id',
 | 
	
		
			
				|  |  | +        tenant_id='tenant_id',
 | 
	
		
			
				|  |  | +        provider_name='huggingface_hub',
 | 
	
		
			
				|  |  | +        provider_type=ProviderType.CUSTOM.value,
 | 
	
		
			
				|  |  | +        encrypted_config='',
 | 
	
		
			
				|  |  | +        is_valid=True,
 | 
	
		
			
				|  |  | +    )
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +def get_mock_embedding_model(model_name, huggingfacehub_api_type, mocker):
 | 
	
		
			
				|  |  | +    valid_api_key = os.environ['HUGGINGFACE_API_KEY']
 | 
	
		
			
				|  |  | +    endpoint_url = os.environ['HUGGINGFACE_EMBEDDINGS_ENDPOINT_URL']
 | 
	
		
			
				|  |  | +    model_provider = HuggingfaceHubProvider(provider=get_mock_provider())
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +    credentials = {
 | 
	
		
			
				|  |  | +        'huggingfacehub_api_type': huggingfacehub_api_type,
 | 
	
		
			
				|  |  | +        'huggingfacehub_api_token': valid_api_key,
 | 
	
		
			
				|  |  | +        'task_type': 'feature-extraction'
 | 
	
		
			
				|  |  | +    }
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +    if huggingfacehub_api_type == 'inference_endpoints':
 | 
	
		
			
				|  |  | +        credentials['huggingfacehub_endpoint_url'] = endpoint_url
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +    mock_query = MagicMock()
 | 
	
		
			
				|  |  | +    mock_query.filter.return_value.first.return_value = ProviderModel(
 | 
	
		
			
				|  |  | +        provider_name='huggingface_hub',
 | 
	
		
			
				|  |  | +        model_name=model_name,
 | 
	
		
			
				|  |  | +        model_type=ModelType.EMBEDDINGS.value,
 | 
	
		
			
				|  |  | +        encrypted_config=json.dumps(credentials),
 | 
	
		
			
				|  |  | +        is_valid=True,
 | 
	
		
			
				|  |  | +    )
 | 
	
		
			
				|  |  | +    mocker.patch('extensions.ext_database.db.session.query',
 | 
	
		
			
				|  |  | +                 return_value=mock_query)
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +    return HuggingfaceEmbedding(
 | 
	
		
			
				|  |  | +        model_provider=model_provider,
 | 
	
		
			
				|  |  | +        name=model_name
 | 
	
		
			
				|  |  | +    )
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +def decrypt_side_effect(tenant_id, encrypted_api_key):
 | 
	
		
			
				|  |  | +    return encrypted_api_key
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +@patch('core.helper.encrypter.decrypt_token', side_effect=decrypt_side_effect)
 | 
	
		
			
				|  |  | +def test_hosted_inference_api_embed_documents(mock_decrypt, mocker):
 | 
	
		
			
				|  |  | +    embedding_model = get_mock_embedding_model(
 | 
	
		
			
				|  |  | +        DEFAULT_MODEL_NAME,
 | 
	
		
			
				|  |  | +        'hosted_inference_api',
 | 
	
		
			
				|  |  | +        mocker)
 | 
	
		
			
				|  |  | +    rst = embedding_model.client.embed_documents(['test', 'test1'])
 | 
	
		
			
				|  |  | +    assert isinstance(rst, list)
 | 
	
		
			
				|  |  | +    assert len(rst) == 2
 | 
	
		
			
				|  |  | +    assert len(rst[0]) == 384
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +@patch('core.helper.encrypter.decrypt_token', side_effect=decrypt_side_effect)
 | 
	
		
			
				|  |  | +def test_endpoint_url_inference_api_embed_documents(mock_decrypt, mocker):
 | 
	
		
			
				|  |  | +    embedding_model = get_mock_embedding_model(
 | 
	
		
			
				|  |  | +        '',
 | 
	
		
			
				|  |  | +        'inference_endpoints',
 | 
	
		
			
				|  |  | +        mocker)
 | 
	
		
			
				|  |  | +    mocker.patch('core.third_party.langchain.embeddings.huggingface_hub_embedding.InferenceClient.post'
 | 
	
		
			
				|  |  | +                 , return_value=bytes(json.dumps([[1, 2, 3], [4, 5, 6]]), 'utf-8'))
 | 
	
		
			
				|  |  | +    
 | 
	
		
			
				|  |  | +    rst = embedding_model.client.embed_documents(['test', 'test1'])
 | 
	
		
			
				|  |  | +    assert isinstance(rst, list)
 | 
	
		
			
				|  |  | +    assert len(rst) == 2
 | 
	
		
			
				|  |  | +    assert len(rst[0]) == 3
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +@patch('core.helper.encrypter.decrypt_token', side_effect=decrypt_side_effect)
 | 
	
		
			
				|  |  | +def test_endpoint_url_inference_api_embed_documents_two(mock_decrypt, mocker):
 | 
	
		
			
				|  |  | +    embedding_model = get_mock_embedding_model(
 | 
	
		
			
				|  |  | +        '',
 | 
	
		
			
				|  |  | +        'inference_endpoints',
 | 
	
		
			
				|  |  | +        mocker)
 | 
	
		
			
				|  |  | +    mocker.patch('core.third_party.langchain.embeddings.huggingface_hub_embedding.InferenceClient.post'
 | 
	
		
			
				|  |  | +                 , return_value=bytes(json.dumps([[[[1,2,3],[4,5,6],[7,8,9]]],[[[1,2,3],[4,5,6],[7,8,9]]]]), 'utf-8'))
 | 
	
		
			
				|  |  | +    
 | 
	
		
			
				|  |  | +    rst = embedding_model.client.embed_documents(['test', 'test1'])
 | 
	
		
			
				|  |  | +    assert isinstance(rst, list)
 | 
	
		
			
				|  |  | +    assert len(rst) == 2
 | 
	
		
			
				|  |  | +    assert len(rst[0]) == 3
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +@patch('core.helper.encrypter.decrypt_token', side_effect=decrypt_side_effect)
 | 
	
		
			
				|  |  | +def test_hosted_inference_api_embed_query(mock_decrypt, mocker):
 | 
	
		
			
				|  |  | +    embedding_model = get_mock_embedding_model(
 | 
	
		
			
				|  |  | +        DEFAULT_MODEL_NAME,
 | 
	
		
			
				|  |  | +        'hosted_inference_api',
 | 
	
		
			
				|  |  | +        mocker)
 | 
	
		
			
				|  |  | +    rst = embedding_model.client.embed_query('test')
 | 
	
		
			
				|  |  | +    assert isinstance(rst, list)
 | 
	
		
			
				|  |  | +    assert len(rst) == 384
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +@patch('core.helper.encrypter.decrypt_token', side_effect=decrypt_side_effect)
 | 
	
		
			
				|  |  | +def test_endpoint_url_inference_api_embed_query(mock_decrypt, mocker):
 | 
	
		
			
				|  |  | +    embedding_model = get_mock_embedding_model(
 | 
	
		
			
				|  |  | +        '',
 | 
	
		
			
				|  |  | +        'inference_endpoints',
 | 
	
		
			
				|  |  | +        mocker)
 | 
	
		
			
				|  |  | +    
 | 
	
		
			
				|  |  | +    mocker.patch('core.third_party.langchain.embeddings.huggingface_hub_embedding.InferenceClient.post'
 | 
	
		
			
				|  |  | +                 , return_value=bytes(json.dumps([[1, 2, 3]]), 'utf-8'))
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +    rst = embedding_model.client.embed_query('test')
 | 
	
		
			
				|  |  | +    assert isinstance(rst, list)
 | 
	
		
			
				|  |  | +    assert len(rst) == 3
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +@patch('core.helper.encrypter.decrypt_token', side_effect=decrypt_side_effect)
 | 
	
		
			
				|  |  | +def test_endpoint_url_inference_api_embed_query_two(mock_decrypt, mocker):
 | 
	
		
			
				|  |  | +    embedding_model = get_mock_embedding_model(
 | 
	
		
			
				|  |  | +        '',
 | 
	
		
			
				|  |  | +        'inference_endpoints',
 | 
	
		
			
				|  |  | +        mocker)
 | 
	
		
			
				|  |  | +    
 | 
	
		
			
				|  |  | +    mocker.patch('core.third_party.langchain.embeddings.huggingface_hub_embedding.InferenceClient.post'
 | 
	
		
			
				|  |  | +                 , return_value=bytes(json.dumps([[[[1,2,3],[4,5,6],[7,8,9]]]]), 'utf-8'))
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +    rst = embedding_model.client.embed_query('test')
 | 
	
		
			
				|  |  | +    assert isinstance(rst, list)
 | 
	
		
			
				|  |  | +    assert len(rst) == 3
 |