test_llm.py 6.7 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215
  1. import os
  2. from collections.abc import Generator
  3. from time import sleep
  4. import pytest
  5. from core.model_runtime.entities.llm_entities import LLMResult, LLMResultChunk, LLMResultChunkDelta
  6. from core.model_runtime.entities.message_entities import AssistantPromptMessage, SystemPromptMessage, UserPromptMessage
  7. from core.model_runtime.entities.model_entities import AIModelEntity
  8. from core.model_runtime.errors.validate import CredentialsValidateFailedError
  9. from core.model_runtime.model_providers.wenxin.llm.llm import ErnieBotLargeLanguageModel
  10. def test_predefined_models():
  11. model = ErnieBotLargeLanguageModel()
  12. model_schemas = model.predefined_models()
  13. assert len(model_schemas) >= 1
  14. assert isinstance(model_schemas[0], AIModelEntity)
  15. def test_validate_credentials_for_chat_model():
  16. sleep(3)
  17. model = ErnieBotLargeLanguageModel()
  18. with pytest.raises(CredentialsValidateFailedError):
  19. model.validate_credentials(
  20. model="ernie-bot", credentials={"api_key": "invalid_key", "secret_key": "invalid_key"}
  21. )
  22. model.validate_credentials(
  23. model="ernie-bot",
  24. credentials={"api_key": os.environ.get("WENXIN_API_KEY"), "secret_key": os.environ.get("WENXIN_SECRET_KEY")},
  25. )
  26. def test_invoke_model_ernie_bot():
  27. sleep(3)
  28. model = ErnieBotLargeLanguageModel()
  29. response = model.invoke(
  30. model="ernie-bot",
  31. credentials={"api_key": os.environ.get("WENXIN_API_KEY"), "secret_key": os.environ.get("WENXIN_SECRET_KEY")},
  32. prompt_messages=[UserPromptMessage(content="Hello World!")],
  33. model_parameters={
  34. "temperature": 0.7,
  35. "top_p": 1.0,
  36. },
  37. stop=["you"],
  38. user="abc-123",
  39. stream=False,
  40. )
  41. assert isinstance(response, LLMResult)
  42. assert len(response.message.content) > 0
  43. assert response.usage.total_tokens > 0
  44. def test_invoke_model_ernie_bot_turbo():
  45. sleep(3)
  46. model = ErnieBotLargeLanguageModel()
  47. response = model.invoke(
  48. model="ernie-bot-turbo",
  49. credentials={"api_key": os.environ.get("WENXIN_API_KEY"), "secret_key": os.environ.get("WENXIN_SECRET_KEY")},
  50. prompt_messages=[UserPromptMessage(content="Hello World!")],
  51. model_parameters={
  52. "temperature": 0.7,
  53. "top_p": 1.0,
  54. },
  55. stop=["you"],
  56. user="abc-123",
  57. stream=False,
  58. )
  59. assert isinstance(response, LLMResult)
  60. assert len(response.message.content) > 0
  61. assert response.usage.total_tokens > 0
  62. def test_invoke_model_ernie_8k():
  63. sleep(3)
  64. model = ErnieBotLargeLanguageModel()
  65. response = model.invoke(
  66. model="ernie-bot-8k",
  67. credentials={"api_key": os.environ.get("WENXIN_API_KEY"), "secret_key": os.environ.get("WENXIN_SECRET_KEY")},
  68. prompt_messages=[UserPromptMessage(content="Hello World!")],
  69. model_parameters={
  70. "temperature": 0.7,
  71. "top_p": 1.0,
  72. },
  73. stop=["you"],
  74. user="abc-123",
  75. stream=False,
  76. )
  77. assert isinstance(response, LLMResult)
  78. assert len(response.message.content) > 0
  79. assert response.usage.total_tokens > 0
  80. def test_invoke_model_ernie_bot_4():
  81. sleep(3)
  82. model = ErnieBotLargeLanguageModel()
  83. response = model.invoke(
  84. model="ernie-bot-4",
  85. credentials={"api_key": os.environ.get("WENXIN_API_KEY"), "secret_key": os.environ.get("WENXIN_SECRET_KEY")},
  86. prompt_messages=[UserPromptMessage(content="Hello World!")],
  87. model_parameters={
  88. "temperature": 0.7,
  89. "top_p": 1.0,
  90. },
  91. stop=["you"],
  92. user="abc-123",
  93. stream=False,
  94. )
  95. assert isinstance(response, LLMResult)
  96. assert len(response.message.content) > 0
  97. assert response.usage.total_tokens > 0
  98. def test_invoke_stream_model():
  99. sleep(3)
  100. model = ErnieBotLargeLanguageModel()
  101. response = model.invoke(
  102. model="ernie-3.5-8k",
  103. credentials={"api_key": os.environ.get("WENXIN_API_KEY"), "secret_key": os.environ.get("WENXIN_SECRET_KEY")},
  104. prompt_messages=[UserPromptMessage(content="Hello World!")],
  105. model_parameters={
  106. "temperature": 0.7,
  107. "top_p": 1.0,
  108. },
  109. stop=["you"],
  110. stream=True,
  111. user="abc-123",
  112. )
  113. assert isinstance(response, Generator)
  114. for chunk in response:
  115. assert isinstance(chunk, LLMResultChunk)
  116. assert isinstance(chunk.delta, LLMResultChunkDelta)
  117. assert isinstance(chunk.delta.message, AssistantPromptMessage)
  118. assert len(chunk.delta.message.content) > 0 if chunk.delta.finish_reason is None else True
  119. def test_invoke_model_with_system():
  120. sleep(3)
  121. model = ErnieBotLargeLanguageModel()
  122. response = model.invoke(
  123. model="ernie-bot",
  124. credentials={"api_key": os.environ.get("WENXIN_API_KEY"), "secret_key": os.environ.get("WENXIN_SECRET_KEY")},
  125. prompt_messages=[SystemPromptMessage(content="你是Kasumi"), UserPromptMessage(content="你是谁?")],
  126. model_parameters={
  127. "temperature": 0.7,
  128. "top_p": 1.0,
  129. },
  130. stop=["you"],
  131. stream=False,
  132. user="abc-123",
  133. )
  134. assert isinstance(response, LLMResult)
  135. assert "kasumi" in response.message.content.lower()
  136. def test_invoke_with_search():
  137. sleep(3)
  138. model = ErnieBotLargeLanguageModel()
  139. response = model.invoke(
  140. model="ernie-bot",
  141. credentials={"api_key": os.environ.get("WENXIN_API_KEY"), "secret_key": os.environ.get("WENXIN_SECRET_KEY")},
  142. prompt_messages=[UserPromptMessage(content="北京今天的天气怎么样")],
  143. model_parameters={
  144. "temperature": 0.7,
  145. "top_p": 1.0,
  146. "disable_search": True,
  147. },
  148. stop=[],
  149. stream=True,
  150. user="abc-123",
  151. )
  152. assert isinstance(response, Generator)
  153. total_message = ""
  154. for chunk in response:
  155. assert isinstance(chunk, LLMResultChunk)
  156. assert isinstance(chunk.delta, LLMResultChunkDelta)
  157. assert isinstance(chunk.delta.message, AssistantPromptMessage)
  158. total_message += chunk.delta.message.content
  159. print(chunk.delta.message.content)
  160. assert len(chunk.delta.message.content) > 0 if not chunk.delta.finish_reason else True
  161. # there should be 对不起、我不能、不支持……
  162. assert "不" in total_message or "抱歉" in total_message or "无法" in total_message
  163. def test_get_num_tokens():
  164. sleep(3)
  165. model = ErnieBotLargeLanguageModel()
  166. response = model.get_num_tokens(
  167. model="ernie-bot",
  168. credentials={"api_key": os.environ.get("WENXIN_API_KEY"), "secret_key": os.environ.get("WENXIN_SECRET_KEY")},
  169. prompt_messages=[UserPromptMessage(content="Hello World!")],
  170. tools=[],
  171. )
  172. assert isinstance(response, int)
  173. assert response == 10