test_xinference_model.py 2.4 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374
  1. import json
  2. import os
  3. from unittest.mock import patch, MagicMock
  4. from core.model_providers.models.entity.message import PromptMessage, MessageType
  5. from core.model_providers.models.entity.model_params import ModelKwargs, ModelType
  6. from core.model_providers.models.llm.xinference_model import XinferenceModel
  7. from core.model_providers.providers.xinference_provider import XinferenceProvider
  8. from models.provider import Provider, ProviderType, ProviderModel
  9. def get_mock_provider():
  10. return Provider(
  11. id='provider_id',
  12. tenant_id='tenant_id',
  13. provider_name='xinference',
  14. provider_type=ProviderType.CUSTOM.value,
  15. encrypted_config='',
  16. is_valid=True,
  17. )
  18. def get_mock_model(model_name, mocker):
  19. model_kwargs = ModelKwargs(
  20. max_tokens=10,
  21. temperature=0.01
  22. )
  23. server_url = os.environ['XINFERENCE_SERVER_URL']
  24. model_uid = os.environ['XINFERENCE_MODEL_UID']
  25. model_provider = XinferenceProvider(provider=get_mock_provider())
  26. mock_query = MagicMock()
  27. mock_query.filter.return_value.first.return_value = ProviderModel(
  28. provider_name='xinference',
  29. model_name=model_name,
  30. model_type=ModelType.TEXT_GENERATION.value,
  31. encrypted_config=json.dumps({
  32. 'server_url': server_url,
  33. 'model_uid': model_uid
  34. }),
  35. is_valid=True,
  36. )
  37. mocker.patch('extensions.ext_database.db.session.query', return_value=mock_query)
  38. return XinferenceModel(
  39. model_provider=model_provider,
  40. name=model_name,
  41. model_kwargs=model_kwargs
  42. )
  43. def decrypt_side_effect(tenant_id, encrypted_api_key):
  44. return encrypted_api_key
  45. @patch('core.helper.encrypter.decrypt_token', side_effect=decrypt_side_effect)
  46. def test_get_num_tokens(mock_decrypt, mocker):
  47. model = get_mock_model('llama-2-chat', mocker)
  48. rst = model.get_num_tokens([
  49. PromptMessage(type=MessageType.HUMAN, content='Who is your manufacturer?')
  50. ])
  51. assert rst == 5
  52. @patch('core.helper.encrypter.decrypt_token', side_effect=decrypt_side_effect)
  53. def test_run(mock_decrypt, mocker):
  54. mocker.patch('core.model_providers.providers.base.BaseModelProvider.update_last_used', return_value=None)
  55. model = get_mock_model('llama-2-chat', mocker)
  56. messages = [PromptMessage(content='Human: 1+1=? \nAnswer: ')]
  57. rst = model.run(
  58. messages
  59. )
  60. assert len(rst.content) > 0