default.ts 2.2 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768
  1. import { BlockEnum } from '../../types'
  2. import { type NodeDefault, PromptRole } from '../../types'
  3. import type { LLMNodeType } from './types'
  4. import type { PromptItem } from '@/models/debug'
  5. import { ALL_CHAT_AVAILABLE_BLOCKS, ALL_COMPLETION_AVAILABLE_BLOCKS } from '@/app/components/workflow/constants'
  6. const i18nPrefix = 'workflow.errorMsg'
  7. const nodeDefault: NodeDefault<LLMNodeType> = {
  8. defaultValue: {
  9. model: {
  10. provider: '',
  11. name: '',
  12. mode: 'chat',
  13. completion_params: {
  14. temperature: 0.7,
  15. },
  16. },
  17. variables: [],
  18. prompt_template: [{
  19. role: PromptRole.system,
  20. text: '',
  21. }],
  22. context: {
  23. enabled: false,
  24. variable_selector: [],
  25. },
  26. vision: {
  27. enabled: false,
  28. },
  29. },
  30. getAvailablePrevNodes(isChatMode: boolean) {
  31. const nodes = isChatMode
  32. ? ALL_CHAT_AVAILABLE_BLOCKS
  33. : ALL_COMPLETION_AVAILABLE_BLOCKS.filter(type => type !== BlockEnum.End)
  34. return nodes
  35. },
  36. getAvailableNextNodes(isChatMode: boolean) {
  37. const nodes = isChatMode ? ALL_CHAT_AVAILABLE_BLOCKS : ALL_COMPLETION_AVAILABLE_BLOCKS
  38. return nodes
  39. },
  40. checkValid(payload: LLMNodeType, t: any) {
  41. let errorMessages = ''
  42. if (!errorMessages && !payload.model.provider)
  43. errorMessages = t(`${i18nPrefix}.fieldRequired`, { field: t(`${i18nPrefix}.fields.model`) })
  44. if (!errorMessages && !payload.memory) {
  45. const isChatModel = payload.model.mode === 'chat'
  46. const isPromptyEmpty = isChatModel ? !(payload.prompt_template as PromptItem[]).some(t => t.text !== '') : (payload.prompt_template as PromptItem).text === ''
  47. if (isPromptyEmpty)
  48. errorMessages = t(`${i18nPrefix}.fieldRequired`, { field: t('workflow.nodes.llm.prompt') })
  49. }
  50. if (!errorMessages && !!payload.memory) {
  51. const isChatModel = payload.model.mode === 'chat'
  52. // payload.memory.query_prompt_template not pass is default: {{#sys.query#}}
  53. if (isChatModel && !!payload.memory.query_prompt_template && !payload.memory.query_prompt_template.includes('{{#sys.query#}}'))
  54. errorMessages = t('workflow.nodes.llm.sysQueryInUser')
  55. }
  56. return {
  57. isValid: !errorMessages,
  58. errorMessage: errorMessages,
  59. }
  60. },
  61. }
  62. export default nodeDefault