|
@@ -489,37 +489,6 @@ if you are not sure about the structure.
|
|
|
"""Cut off the text as soon as any stop words occur."""
|
|
|
return re.split("|".join(stop), text, maxsplit=1)[0]
|
|
|
|
|
|
- def _llm_result_to_stream(self, result: LLMResult) -> Generator:
|
|
|
- """
|
|
|
-from typing_extensions import deprecated
|
|
|
- Transform llm result to stream
|
|
|
-
|
|
|
- :param result: llm result
|
|
|
- :return: stream
|
|
|
- """
|
|
|
- index = 0
|
|
|
-
|
|
|
- tool_calls = result.message.tool_calls
|
|
|
-
|
|
|
- for word in result.message.content:
|
|
|
- assistant_prompt_message = AssistantPromptMessage(
|
|
|
- content=word,
|
|
|
- tool_calls=tool_calls if index == (len(result.message.content) - 1) else []
|
|
|
- )
|
|
|
-
|
|
|
- yield LLMResultChunk(
|
|
|
- model=result.model,
|
|
|
- prompt_messages=result.prompt_messages,
|
|
|
- system_fingerprint=result.system_fingerprint,
|
|
|
- delta=LLMResultChunkDelta(
|
|
|
- index=index,
|
|
|
- message=assistant_prompt_message,
|
|
|
- )
|
|
|
- )
|
|
|
-
|
|
|
- index += 1
|
|
|
- time.sleep(0.01)
|
|
|
-
|
|
|
def get_parameter_rules(self, model: str, credentials: dict) -> list[ParameterRule]:
|
|
|
"""
|
|
|
Get parameter rules
|