conversation_message_task.py 17 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412
  1. import decimal
  2. import json
  3. from typing import Optional, Union
  4. from core.callback_handler.entity.agent_loop import AgentLoop
  5. from core.callback_handler.entity.dataset_query import DatasetQueryObj
  6. from core.callback_handler.entity.llm_message import LLMMessage
  7. from core.callback_handler.entity.chain_result import ChainResult
  8. from core.constant import llm_constant
  9. from core.llm.llm_builder import LLMBuilder
  10. from core.llm.provider.llm_provider_service import LLMProviderService
  11. from core.prompt.prompt_builder import PromptBuilder
  12. from core.prompt.prompt_template import JinjaPromptTemplate
  13. from events.message_event import message_was_created
  14. from extensions.ext_database import db
  15. from extensions.ext_redis import redis_client
  16. from models.dataset import DatasetQuery
  17. from models.model import AppModelConfig, Conversation, Account, Message, EndUser, App, MessageAgentThought, MessageChain
  18. from models.provider import ProviderType, Provider
  19. class ConversationMessageTask:
  20. def __init__(self, task_id: str, app: App, app_model_config: AppModelConfig, user: Account,
  21. inputs: dict, query: str, streaming: bool,
  22. conversation: Optional[Conversation] = None, is_override: bool = False):
  23. self.task_id = task_id
  24. self.app = app
  25. self.tenant_id = app.tenant_id
  26. self.app_model_config = app_model_config
  27. self.is_override = is_override
  28. self.user = user
  29. self.inputs = inputs
  30. self.query = query
  31. self.streaming = streaming
  32. self.conversation = conversation
  33. self.is_new_conversation = False
  34. self.message = None
  35. self.model_dict = self.app_model_config.model_dict
  36. self.model_name = self.model_dict.get('name')
  37. self.mode = app.mode
  38. self.init()
  39. self._pub_handler = PubHandler(
  40. user=self.user,
  41. task_id=self.task_id,
  42. message=self.message,
  43. conversation=self.conversation,
  44. chain_pub=False, # disabled currently
  45. agent_thought_pub=True
  46. )
  47. def init(self):
  48. provider_name = LLMBuilder.get_default_provider(self.app.tenant_id, self.model_name)
  49. self.model_dict['provider'] = provider_name
  50. override_model_configs = None
  51. if self.is_override:
  52. override_model_configs = {
  53. "model": self.app_model_config.model_dict,
  54. "pre_prompt": self.app_model_config.pre_prompt,
  55. "agent_mode": self.app_model_config.agent_mode_dict,
  56. "opening_statement": self.app_model_config.opening_statement,
  57. "suggested_questions": self.app_model_config.suggested_questions_list,
  58. "suggested_questions_after_answer": self.app_model_config.suggested_questions_after_answer_dict,
  59. "more_like_this": self.app_model_config.more_like_this_dict,
  60. "sensitive_word_avoidance": self.app_model_config.sensitive_word_avoidance_dict,
  61. "user_input_form": self.app_model_config.user_input_form_list,
  62. }
  63. introduction = ''
  64. system_instruction = ''
  65. system_instruction_tokens = 0
  66. if self.mode == 'chat':
  67. introduction = self.app_model_config.opening_statement
  68. if introduction:
  69. prompt_template = JinjaPromptTemplate.from_template(template=introduction)
  70. prompt_inputs = {k: self.inputs[k] for k in prompt_template.input_variables if k in self.inputs}
  71. try:
  72. introduction = prompt_template.format(**prompt_inputs)
  73. except KeyError:
  74. pass
  75. if self.app_model_config.pre_prompt:
  76. system_message = PromptBuilder.to_system_message(self.app_model_config.pre_prompt, self.inputs)
  77. system_instruction = system_message.content
  78. llm = LLMBuilder.to_llm(self.tenant_id, self.model_name)
  79. system_instruction_tokens = llm.get_num_tokens_from_messages([system_message])
  80. if not self.conversation:
  81. self.is_new_conversation = True
  82. self.conversation = Conversation(
  83. app_id=self.app_model_config.app_id,
  84. app_model_config_id=self.app_model_config.id,
  85. model_provider=self.model_dict.get('provider'),
  86. model_id=self.model_name,
  87. override_model_configs=json.dumps(override_model_configs) if override_model_configs else None,
  88. mode=self.mode,
  89. name='',
  90. inputs=self.inputs,
  91. introduction=introduction,
  92. system_instruction=system_instruction,
  93. system_instruction_tokens=system_instruction_tokens,
  94. status='normal',
  95. from_source=('console' if isinstance(self.user, Account) else 'api'),
  96. from_end_user_id=(self.user.id if isinstance(self.user, EndUser) else None),
  97. from_account_id=(self.user.id if isinstance(self.user, Account) else None),
  98. )
  99. db.session.add(self.conversation)
  100. db.session.flush()
  101. self.message = Message(
  102. app_id=self.app_model_config.app_id,
  103. model_provider=self.model_dict.get('provider'),
  104. model_id=self.model_name,
  105. override_model_configs=json.dumps(override_model_configs) if override_model_configs else None,
  106. conversation_id=self.conversation.id,
  107. inputs=self.inputs,
  108. query=self.query,
  109. message="",
  110. message_tokens=0,
  111. message_unit_price=0,
  112. answer="",
  113. answer_tokens=0,
  114. answer_unit_price=0,
  115. provider_response_latency=0,
  116. total_price=0,
  117. currency=llm_constant.model_currency,
  118. from_source=('console' if isinstance(self.user, Account) else 'api'),
  119. from_end_user_id=(self.user.id if isinstance(self.user, EndUser) else None),
  120. from_account_id=(self.user.id if isinstance(self.user, Account) else None),
  121. agent_based=self.app_model_config.agent_mode_dict.get('enabled'),
  122. )
  123. db.session.add(self.message)
  124. db.session.flush()
  125. def append_message_text(self, text: str):
  126. self._pub_handler.pub_text(text)
  127. def save_message(self, llm_message: LLMMessage, by_stopped: bool = False):
  128. model_name = self.app_model_config.model_dict.get('name')
  129. message_tokens = llm_message.prompt_tokens
  130. answer_tokens = llm_message.completion_tokens
  131. message_unit_price = llm_constant.model_prices[model_name]['prompt']
  132. answer_unit_price = llm_constant.model_prices[model_name]['completion']
  133. total_price = self.calc_total_price(message_tokens, message_unit_price, answer_tokens, answer_unit_price)
  134. self.message.message = llm_message.prompt
  135. self.message.message_tokens = message_tokens
  136. self.message.message_unit_price = message_unit_price
  137. self.message.answer = PromptBuilder.process_template(llm_message.completion.strip()) if llm_message.completion else ''
  138. self.message.answer_tokens = answer_tokens
  139. self.message.answer_unit_price = answer_unit_price
  140. self.message.provider_response_latency = llm_message.latency
  141. self.message.total_price = total_price
  142. self.update_provider_quota()
  143. db.session.commit()
  144. message_was_created.send(
  145. self.message,
  146. conversation=self.conversation,
  147. is_first_message=self.is_new_conversation
  148. )
  149. if not by_stopped:
  150. self.end()
  151. def update_provider_quota(self):
  152. llm_provider_service = LLMProviderService(
  153. tenant_id=self.app.tenant_id,
  154. provider_name=self.message.model_provider,
  155. )
  156. provider = llm_provider_service.get_provider_db_record()
  157. if provider and provider.provider_type == ProviderType.SYSTEM.value:
  158. db.session.query(Provider).filter(
  159. Provider.tenant_id == self.app.tenant_id,
  160. Provider.provider_name == provider.provider_name,
  161. Provider.quota_limit > Provider.quota_used
  162. ).update({'quota_used': Provider.quota_used + 1})
  163. def init_chain(self, chain_result: ChainResult):
  164. message_chain = MessageChain(
  165. message_id=self.message.id,
  166. type=chain_result.type,
  167. input=json.dumps(chain_result.prompt),
  168. output=''
  169. )
  170. db.session.add(message_chain)
  171. db.session.flush()
  172. return message_chain
  173. def on_chain_end(self, message_chain: MessageChain, chain_result: ChainResult):
  174. message_chain.output = json.dumps(chain_result.completion)
  175. self._pub_handler.pub_chain(message_chain)
  176. def on_agent_start(self, message_chain: MessageChain, agent_loop: AgentLoop) -> MessageAgentThought:
  177. message_agent_thought = MessageAgentThought(
  178. message_id=self.message.id,
  179. message_chain_id=message_chain.id,
  180. position=agent_loop.position,
  181. thought=agent_loop.thought,
  182. tool=agent_loop.tool_name,
  183. tool_input=agent_loop.tool_input,
  184. message=agent_loop.prompt,
  185. answer=agent_loop.completion,
  186. created_by_role=('account' if isinstance(self.user, Account) else 'end_user'),
  187. created_by=self.user.id
  188. )
  189. db.session.add(message_agent_thought)
  190. db.session.flush()
  191. self._pub_handler.pub_agent_thought(message_agent_thought)
  192. return message_agent_thought
  193. def on_agent_end(self, message_agent_thought: MessageAgentThought, agent_model_name: str,
  194. agent_loop: AgentLoop):
  195. agent_message_unit_price = llm_constant.model_prices[agent_model_name]['prompt']
  196. agent_answer_unit_price = llm_constant.model_prices[agent_model_name]['completion']
  197. loop_message_tokens = agent_loop.prompt_tokens
  198. loop_answer_tokens = agent_loop.completion_tokens
  199. loop_total_price = self.calc_total_price(
  200. loop_message_tokens,
  201. agent_message_unit_price,
  202. loop_answer_tokens,
  203. agent_answer_unit_price
  204. )
  205. message_agent_thought.observation = agent_loop.tool_output
  206. message_agent_thought.tool_process_data = '' # currently not support
  207. message_agent_thought.message_token = loop_message_tokens
  208. message_agent_thought.message_unit_price = agent_message_unit_price
  209. message_agent_thought.answer_token = loop_answer_tokens
  210. message_agent_thought.answer_unit_price = agent_answer_unit_price
  211. message_agent_thought.latency = agent_loop.latency
  212. message_agent_thought.tokens = agent_loop.prompt_tokens + agent_loop.completion_tokens
  213. message_agent_thought.total_price = loop_total_price
  214. message_agent_thought.currency = llm_constant.model_currency
  215. db.session.flush()
  216. def on_dataset_query_end(self, dataset_query_obj: DatasetQueryObj):
  217. dataset_query = DatasetQuery(
  218. dataset_id=dataset_query_obj.dataset_id,
  219. content=dataset_query_obj.query,
  220. source='app',
  221. source_app_id=self.app.id,
  222. created_by_role=('account' if isinstance(self.user, Account) else 'end_user'),
  223. created_by=self.user.id
  224. )
  225. db.session.add(dataset_query)
  226. def calc_total_price(self, message_tokens, message_unit_price, answer_tokens, answer_unit_price):
  227. message_tokens_per_1k = (decimal.Decimal(message_tokens) / 1000).quantize(decimal.Decimal('0.001'),
  228. rounding=decimal.ROUND_HALF_UP)
  229. answer_tokens_per_1k = (decimal.Decimal(answer_tokens) / 1000).quantize(decimal.Decimal('0.001'),
  230. rounding=decimal.ROUND_HALF_UP)
  231. total_price = message_tokens_per_1k * message_unit_price + answer_tokens_per_1k * answer_unit_price
  232. return total_price.quantize(decimal.Decimal('0.0000001'), rounding=decimal.ROUND_HALF_UP)
  233. def end(self):
  234. self._pub_handler.pub_end()
  235. class PubHandler:
  236. def __init__(self, user: Union[Account | EndUser], task_id: str,
  237. message: Message, conversation: Conversation,
  238. chain_pub: bool = False, agent_thought_pub: bool = False):
  239. self._channel = PubHandler.generate_channel_name(user, task_id)
  240. self._stopped_cache_key = PubHandler.generate_stopped_cache_key(user, task_id)
  241. self._task_id = task_id
  242. self._message = message
  243. self._conversation = conversation
  244. self._chain_pub = chain_pub
  245. self._agent_thought_pub = agent_thought_pub
  246. @classmethod
  247. def generate_channel_name(cls, user: Union[Account | EndUser], task_id: str):
  248. if not user:
  249. raise ValueError("user is required")
  250. user_str = 'account-' + str(user.id) if isinstance(user, Account) else 'end-user-' + str(user.id)
  251. return "generate_result:{}-{}".format(user_str, task_id)
  252. @classmethod
  253. def generate_stopped_cache_key(cls, user: Union[Account | EndUser], task_id: str):
  254. user_str = 'account-' + str(user.id) if isinstance(user, Account) else 'end-user-' + str(user.id)
  255. return "generate_result_stopped:{}-{}".format(user_str, task_id)
  256. def pub_text(self, text: str):
  257. content = {
  258. 'event': 'message',
  259. 'data': {
  260. 'task_id': self._task_id,
  261. 'message_id': str(self._message.id),
  262. 'text': text,
  263. 'mode': self._conversation.mode,
  264. 'conversation_id': str(self._conversation.id)
  265. }
  266. }
  267. redis_client.publish(self._channel, json.dumps(content))
  268. if self._is_stopped():
  269. self.pub_end()
  270. raise ConversationTaskStoppedException()
  271. def pub_chain(self, message_chain: MessageChain):
  272. if self._chain_pub:
  273. content = {
  274. 'event': 'chain',
  275. 'data': {
  276. 'task_id': self._task_id,
  277. 'message_id': self._message.id,
  278. 'chain_id': message_chain.id,
  279. 'type': message_chain.type,
  280. 'input': json.loads(message_chain.input),
  281. 'output': json.loads(message_chain.output),
  282. 'mode': self._conversation.mode,
  283. 'conversation_id': self._conversation.id
  284. }
  285. }
  286. redis_client.publish(self._channel, json.dumps(content))
  287. if self._is_stopped():
  288. self.pub_end()
  289. raise ConversationTaskStoppedException()
  290. def pub_agent_thought(self, message_agent_thought: MessageAgentThought):
  291. if self._agent_thought_pub:
  292. content = {
  293. 'event': 'agent_thought',
  294. 'data': {
  295. 'id': message_agent_thought.id,
  296. 'task_id': self._task_id,
  297. 'message_id': self._message.id,
  298. 'chain_id': message_agent_thought.message_chain_id,
  299. 'position': message_agent_thought.position,
  300. 'thought': message_agent_thought.thought,
  301. 'tool': message_agent_thought.tool,
  302. 'tool_input': message_agent_thought.tool_input,
  303. 'mode': self._conversation.mode,
  304. 'conversation_id': self._conversation.id
  305. }
  306. }
  307. redis_client.publish(self._channel, json.dumps(content))
  308. if self._is_stopped():
  309. self.pub_end()
  310. raise ConversationTaskStoppedException()
  311. def pub_end(self):
  312. content = {
  313. 'event': 'end',
  314. }
  315. redis_client.publish(self._channel, json.dumps(content))
  316. @classmethod
  317. def pub_error(cls, user: Union[Account | EndUser], task_id: str, e):
  318. content = {
  319. 'error': type(e).__name__,
  320. 'description': e.description if getattr(e, 'description', None) is not None else str(e)
  321. }
  322. channel = cls.generate_channel_name(user, task_id)
  323. redis_client.publish(channel, json.dumps(content))
  324. def _is_stopped(self):
  325. return redis_client.get(self._stopped_cache_key) is not None
  326. @classmethod
  327. def ping(cls, user: Union[Account | EndUser], task_id: str):
  328. content = {
  329. 'event': 'ping'
  330. }
  331. channel = cls.generate_channel_name(user, task_id)
  332. redis_client.publish(channel, json.dumps(content))
  333. @classmethod
  334. def stop(cls, user: Union[Account | EndUser], task_id: str):
  335. stopped_cache_key = cls.generate_stopped_cache_key(user, task_id)
  336. redis_client.setex(stopped_cache_key, 600, 1)
  337. class ConversationTaskStoppedException(Exception):
  338. pass