Browse Source

Update Novita AI predefined models (#13045)

Jason 2 months ago
parent
commit
d4be5ef9de
25 changed files with 614 additions and 56 deletions
  1. 8 16
      api/core/model_runtime/model_providers/novita/_assets/icon_l_en.svg
  2. 2 9
      api/core/model_runtime/model_providers/novita/_assets/icon_s_en.svg
  3. 41 0
      api/core/model_runtime/model_providers/novita/llm/L3-8B-Stheno-v3.2.yaml
  4. 40 0
      api/core/model_runtime/model_providers/novita/llm/_position.yaml
  5. 41 0
      api/core/model_runtime/model_providers/novita/llm/deepseek_v3.yaml
  6. 41 0
      api/core/model_runtime/model_providers/novita/llm/l3-8b-lunaris.yaml
  7. 41 0
      api/core/model_runtime/model_providers/novita/llm/l31-70b-euryale-v2.2.yaml
  8. 2 2
      api/core/model_runtime/model_providers/novita/llm/llama-3-8b-instruct.yaml
  9. 3 3
      api/core/model_runtime/model_providers/novita/llm/llama-3.1-70b-instruct.yaml
  10. 41 0
      api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct-bf16.yaml
  11. 41 0
      api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct-max.yaml
  12. 3 3
      api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct.yaml
  13. 41 0
      api/core/model_runtime/model_providers/novita/llm/llama-3.2-11b-vision-instruct.yaml
  14. 41 0
      api/core/model_runtime/model_providers/novita/llm/llama-3.2-1b-instruct.yaml
  15. 41 0
      api/core/model_runtime/model_providers/novita/llm/llama-3.2-3b-instruct.yaml
  16. 41 0
      api/core/model_runtime/model_providers/novita/llm/llama-3.3-70b-instruct.yaml
  17. 41 0
      api/core/model_runtime/model_providers/novita/llm/mistral-nemo.yaml
  18. 2 2
      api/core/model_runtime/model_providers/novita/llm/mythomax-l2-13b.yaml
  19. 5 5
      api/core/model_runtime/model_providers/novita/llm/lzlv_70b.yaml
  20. 5 5
      api/core/model_runtime/model_providers/novita/llm/Nous-Hermes-2-Mixtral-8x7B-DPO.yaml
  21. 5 5
      api/core/model_runtime/model_providers/novita/llm/llama-3.1-405b-instruct.yaml
  22. 41 0
      api/core/model_runtime/model_providers/novita/llm/qwen-2-vl-72b-instruct.yaml
  23. 41 0
      api/core/model_runtime/model_providers/novita/llm/qwen-2.5-72b-instruct.yaml
  24. 2 2
      api/core/model_runtime/model_providers/novita/llm/wizardlm-2-8x22b.yaml
  25. 4 4
      api/core/model_runtime/model_providers/novita/novita.yaml

File diff suppressed because it is too large
+ 8 - 16
api/core/model_runtime/model_providers/novita/_assets/icon_l_en.svg


File diff suppressed because it is too large
+ 2 - 9
api/core/model_runtime/model_providers/novita/_assets/icon_s_en.svg


+ 41 - 0
api/core/model_runtime/model_providers/novita/llm/L3-8B-Stheno-v3.2.yaml

@@ -0,0 +1,41 @@
+model: Sao10K/L3-8B-Stheno-v3.2
+label:
+  zh_Hans: Sao10K/L3-8B-Stheno-v3.2
+  en_US: Sao10K/L3-8B-Stheno-v3.2
+model_type: llm
+features:
+  - agent-thought
+model_properties:
+  mode: chat
+  context_size: 8192
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+    min: 0
+    max: 2
+    default: 1
+  - name: top_p
+    use_template: top_p
+    min: 0
+    max: 1
+    default: 1
+  - name: max_tokens
+    use_template: max_tokens
+    min: 1
+    max: 2048
+    default: 512
+  - name: frequency_penalty
+    use_template: frequency_penalty
+    min: -2
+    max: 2
+    default: 0
+  - name: presence_penalty
+    use_template: presence_penalty
+    min: -2
+    max: 2
+    default: 0
+pricing:
+  input: '0.0005'
+  output: '0.0005'
+  unit: '0.0001'
+  currency: USD

+ 40 - 0
api/core/model_runtime/model_providers/novita/llm/_position.yaml

@@ -0,0 +1,40 @@
+# Deepseek Models
+- deepseek/deepseek_v3
+
+# LLaMA Models
+- meta-llama/llama-3.3-70b-instruct
+- meta-llama/llama-3.2-11b-vision-instruct
+- meta-llama/llama-3.2-3b-instruct
+- meta-llama/llama-3.2-1b-instruct
+- meta-llama/llama-3.1-70b-instruct
+- meta-llama/llama-3.1-8b-instruct
+- meta-llama/llama-3.1-8b-instruct-max
+- meta-llama/llama-3.1-8b-instruct-bf16
+- meta-llama/llama-3-70b-instruct
+- meta-llama/llama-3-8b-instruct
+
+# Mistral Models
+- mistralai/mistral-nemo
+- mistralai/mistral-7b-instruct
+
+# Qwen Models
+- qwen/qwen-2.5-72b-instruct
+- qwen/qwen-2-72b-instruct
+- qwen/qwen-2-vl-72b-instruct
+- qwen/qwen-2-7b-instruct
+
+# Other Models
+- sao10k/L3-8B-Stheno-v3.2
+- sao10k/l3-70b-euryale-v2.1
+- sao10k/l31-70b-euryale-v2.2
+- sao10k/l3-8b-lunaris
+- jondurbin/airoboros-l2-70b
+- cognitivecomputations/dolphin-mixtral-8x22b
+- google/gemma-2-9b-it
+- nousresearch/hermes-2-pro-llama-3-8b
+- sophosympatheia/midnight-rose-70b
+- gryphe/mythomax-l2-13b
+- nousresearch/nous-hermes-llama2-13b
+- openchat/openchat-7b
+- teknium/openhermes-2.5-mistral-7b
+- microsoft/wizardlm-2-8x22b

+ 41 - 0
api/core/model_runtime/model_providers/novita/llm/deepseek_v3.yaml

@@ -0,0 +1,41 @@
+model: deepseek/deepseek_v3
+label:
+  zh_Hans: deepseek/deepseek_v3
+  en_US: deepseek/deepseek_v3
+model_type: llm
+features:
+  - agent-thought
+model_properties:
+  mode: chat
+  context_size: 64000
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+    min: 0
+    max: 2
+    default: 1
+  - name: top_p
+    use_template: top_p
+    min: 0
+    max: 1
+    default: 1
+  - name: max_tokens
+    use_template: max_tokens
+    min: 1
+    max: 2048
+    default: 512
+  - name: frequency_penalty
+    use_template: frequency_penalty
+    min: -2
+    max: 2
+    default: 0
+  - name: presence_penalty
+    use_template: presence_penalty
+    min: -2
+    max: 2
+    default: 0
+pricing:
+  input: '0.0089'
+  output: '0.0089'
+  unit: '0.0001'
+  currency: USD

+ 41 - 0
api/core/model_runtime/model_providers/novita/llm/l3-8b-lunaris.yaml

@@ -0,0 +1,41 @@
+model: sao10k/l3-8b-lunaris
+label:
+  zh_Hans: sao10k/l3-8b-lunaris
+  en_US: sao10k/l3-8b-lunaris
+model_type: llm
+features:
+  - agent-thought
+model_properties:
+  mode: chat
+  context_size: 8192
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+    min: 0
+    max: 2
+    default: 1
+  - name: top_p
+    use_template: top_p
+    min: 0
+    max: 1
+    default: 1
+  - name: max_tokens
+    use_template: max_tokens
+    min: 1
+    max: 2048
+    default: 512
+  - name: frequency_penalty
+    use_template: frequency_penalty
+    min: -2
+    max: 2
+    default: 0
+  - name: presence_penalty
+    use_template: presence_penalty
+    min: -2
+    max: 2
+    default: 0
+pricing:
+  input: '0.0005'
+  output: '0.0005'
+  unit: '0.0001'
+  currency: USD

+ 41 - 0
api/core/model_runtime/model_providers/novita/llm/l31-70b-euryale-v2.2.yaml

@@ -0,0 +1,41 @@
+model: sao10k/l31-70b-euryale-v2.2
+label:
+  zh_Hans: sao10k/l31-70b-euryale-v2.2
+  en_US: sao10k/l31-70b-euryale-v2.2
+model_type: llm
+features:
+  - agent-thought
+model_properties:
+  mode: chat
+  context_size: 16000
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+    min: 0
+    max: 2
+    default: 1
+  - name: top_p
+    use_template: top_p
+    min: 0
+    max: 1
+    default: 1
+  - name: max_tokens
+    use_template: max_tokens
+    min: 1
+    max: 2048
+    default: 512
+  - name: frequency_penalty
+    use_template: frequency_penalty
+    min: -2
+    max: 2
+    default: 0
+  - name: presence_penalty
+    use_template: presence_penalty
+    min: -2
+    max: 2
+    default: 0
+pricing:
+  input: '0.0148'
+  output: '0.0148'
+  unit: '0.0001'
+  currency: USD

+ 2 - 2
api/core/model_runtime/model_providers/novita/llm/llama-3-8b-instruct.yaml

@@ -35,7 +35,7 @@ parameter_rules:
     max: 2
     default: 0
 pricing:
-  input: '0.00063'
-  output: '0.00063'
+  input: '0.0004'
+  output: '0.0004'
   unit: '0.0001'
   currency: USD

+ 3 - 3
api/core/model_runtime/model_providers/novita/llm/llama-3.1-70b-instruct.yaml

@@ -7,7 +7,7 @@ features:
   - agent-thought
 model_properties:
   mode: chat
-  context_size: 8192
+  context_size: 32768
 parameter_rules:
   - name: temperature
     use_template: temperature
@@ -35,7 +35,7 @@ parameter_rules:
     max: 2
     default: 0
 pricing:
-  input: '0.0055'
-  output: '0.0076'
+  input: '0.0034'
+  output: '0.0039'
   unit: '0.0001'
   currency: USD

+ 41 - 0
api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct-bf16.yaml

@@ -0,0 +1,41 @@
+model: meta-llama/llama-3.1-8b-instruct-bf16
+label:
+  zh_Hans: meta-llama/llama-3.1-8b-instruct-bf16
+  en_US: meta-llama/llama-3.1-8b-instruct-bf16
+model_type: llm
+features:
+  - agent-thought
+model_properties:
+  mode: chat
+  context_size: 8192
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+    min: 0
+    max: 2
+    default: 1
+  - name: top_p
+    use_template: top_p
+    min: 0
+    max: 1
+    default: 1
+  - name: max_tokens
+    use_template: max_tokens
+    min: 1
+    max: 2048
+    default: 512
+  - name: frequency_penalty
+    use_template: frequency_penalty
+    min: -2
+    max: 2
+    default: 0
+  - name: presence_penalty
+    use_template: presence_penalty
+    min: -2
+    max: 2
+    default: 0
+pricing:
+  input: '0.0006'
+  output: '0.0006'
+  unit: '0.0001'
+  currency: USD

+ 41 - 0
api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct-max.yaml

@@ -0,0 +1,41 @@
+model: meta-llama/llama-3.1-8b-instruct-max
+label:
+  zh_Hans: meta-llama/llama-3.1-8b-instruct-max
+  en_US: meta-llama/llama-3.1-8b-instruct-max
+model_type: llm
+features:
+  - agent-thought
+model_properties:
+  mode: chat
+  context_size: 16384
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+    min: 0
+    max: 2
+    default: 1
+  - name: top_p
+    use_template: top_p
+    min: 0
+    max: 1
+    default: 1
+  - name: max_tokens
+    use_template: max_tokens
+    min: 1
+    max: 2048
+    default: 512
+  - name: frequency_penalty
+    use_template: frequency_penalty
+    min: -2
+    max: 2
+    default: 0
+  - name: presence_penalty
+    use_template: presence_penalty
+    min: -2
+    max: 2
+    default: 0
+pricing:
+  input: '0.0005'
+  output: '0.0005'
+  unit: '0.0001'
+  currency: USD

+ 3 - 3
api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct.yaml

@@ -7,7 +7,7 @@ features:
   - agent-thought
 model_properties:
   mode: chat
-  context_size: 8192
+  context_size: 16384
 parameter_rules:
   - name: temperature
     use_template: temperature
@@ -35,7 +35,7 @@ parameter_rules:
     max: 2
     default: 0
 pricing:
-  input: '0.001'
-  output: '0.001'
+  input: '0.0005'
+  output: '0.0005'
   unit: '0.0001'
   currency: USD

+ 41 - 0
api/core/model_runtime/model_providers/novita/llm/llama-3.2-11b-vision-instruct.yaml

@@ -0,0 +1,41 @@
+model: meta-llama/llama-3.2-11b-vision-instruct
+label:
+  zh_Hans: meta-llama/llama-3.2-11b-vision-instruct
+  en_US: meta-llama/llama-3.2-11b-vision-instruct
+model_type: llm
+features:
+  - agent-thought
+model_properties:
+  mode: chat
+  context_size: 32768
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+    min: 0
+    max: 2
+    default: 1
+  - name: top_p
+    use_template: top_p
+    min: 0
+    max: 1
+    default: 1
+  - name: max_tokens
+    use_template: max_tokens
+    min: 1
+    max: 2048
+    default: 512
+  - name: frequency_penalty
+    use_template: frequency_penalty
+    min: -2
+    max: 2
+    default: 0
+  - name: presence_penalty
+    use_template: presence_penalty
+    min: -2
+    max: 2
+    default: 0
+pricing:
+  input: '0.0006'
+  output: '0.0006'
+  unit: '0.0001'
+  currency: USD

+ 41 - 0
api/core/model_runtime/model_providers/novita/llm/llama-3.2-1b-instruct.yaml

@@ -0,0 +1,41 @@
+model: meta-llama/llama-3.2-1b-instruct
+label:
+  zh_Hans: meta-llama/llama-3.2-1b-instruct
+  en_US: meta-llama/llama-3.2-1b-instruct
+model_type: llm
+features:
+  - agent-thought
+model_properties:
+  mode: chat
+  context_size: 131000
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+    min: 0
+    max: 2
+    default: 1
+  - name: top_p
+    use_template: top_p
+    min: 0
+    max: 1
+    default: 1
+  - name: max_tokens
+    use_template: max_tokens
+    min: 1
+    max: 2048
+    default: 512
+  - name: frequency_penalty
+    use_template: frequency_penalty
+    min: -2
+    max: 2
+    default: 0
+  - name: presence_penalty
+    use_template: presence_penalty
+    min: -2
+    max: 2
+    default: 0
+pricing:
+  input: '0.0002'
+  output: '0.0002'
+  unit: '0.0001'
+  currency: USD

+ 41 - 0
api/core/model_runtime/model_providers/novita/llm/llama-3.2-3b-instruct.yaml

@@ -0,0 +1,41 @@
+model: meta-llama/llama-3.2-3b-instruct
+label:
+  zh_Hans: meta-llama/llama-3.2-3b-instruct
+  en_US: meta-llama/llama-3.2-3b-instruct
+model_type: llm
+features:
+  - agent-thought
+model_properties:
+  mode: chat
+  context_size: 32768
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+    min: 0
+    max: 2
+    default: 1
+  - name: top_p
+    use_template: top_p
+    min: 0
+    max: 1
+    default: 1
+  - name: max_tokens
+    use_template: max_tokens
+    min: 1
+    max: 2048
+    default: 512
+  - name: frequency_penalty
+    use_template: frequency_penalty
+    min: -2
+    max: 2
+    default: 0
+  - name: presence_penalty
+    use_template: presence_penalty
+    min: -2
+    max: 2
+    default: 0
+pricing:
+  input: '0.0003'
+  output: '0.0005'
+  unit: '0.0001'
+  currency: USD

+ 41 - 0
api/core/model_runtime/model_providers/novita/llm/llama-3.3-70b-instruct.yaml

@@ -0,0 +1,41 @@
+model: meta-llama/llama-3.3-70b-instruct
+label:
+  zh_Hans: meta-llama/llama-3.3-70b-instruct
+  en_US: meta-llama/llama-3.3-70b-instruct
+model_type: llm
+features:
+  - agent-thought
+model_properties:
+  mode: chat
+  context_size: 131072
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+    min: 0
+    max: 2
+    default: 1
+  - name: top_p
+    use_template: top_p
+    min: 0
+    max: 1
+    default: 1
+  - name: max_tokens
+    use_template: max_tokens
+    min: 1
+    max: 2048
+    default: 512
+  - name: frequency_penalty
+    use_template: frequency_penalty
+    min: -2
+    max: 2
+    default: 0
+  - name: presence_penalty
+    use_template: presence_penalty
+    min: -2
+    max: 2
+    default: 0
+pricing:
+  input: '0.0039'
+  output: '0.0039'
+  unit: '0.0001'
+  currency: USD

+ 41 - 0
api/core/model_runtime/model_providers/novita/llm/mistral-nemo.yaml

@@ -0,0 +1,41 @@
+model: mistralai/mistral-nemo
+label:
+  zh_Hans: mistralai/mistral-nemo
+  en_US: mistralai/mistral-nemo
+model_type: llm
+features:
+  - agent-thought
+model_properties:
+  mode: chat
+  context_size: 131072
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+    min: 0
+    max: 2
+    default: 1
+  - name: top_p
+    use_template: top_p
+    min: 0
+    max: 1
+    default: 1
+  - name: max_tokens
+    use_template: max_tokens
+    min: 1
+    max: 2048
+    default: 512
+  - name: frequency_penalty
+    use_template: frequency_penalty
+    min: -2
+    max: 2
+    default: 0
+  - name: presence_penalty
+    use_template: presence_penalty
+    min: -2
+    max: 2
+    default: 0
+pricing:
+  input: '0.0017'
+  output: '0.0017'
+  unit: '0.0001'
+  currency: USD

+ 2 - 2
api/core/model_runtime/model_providers/novita/llm/mythomax-l2-13b.yaml

@@ -35,7 +35,7 @@ parameter_rules:
     max: 2
     default: 0
 pricing:
-  input: '0.00119'
-  output: '0.00119'
+  input: '0.0009'
+  output: '0.0009'
   unit: '0.0001'
   currency: USD

+ 5 - 5
api/core/model_runtime/model_providers/novita/llm/lzlv_70b.yaml

@@ -1,7 +1,7 @@
-model: lzlv_70b
+model: openchat/openchat-7b
 label:
-  zh_Hans: lzlv_70b
-  en_US: lzlv_70b
+  zh_Hans: openchat/openchat-7b
+  en_US: openchat/openchat-7b
 model_type: llm
 features:
   - agent-thought
@@ -35,7 +35,7 @@ parameter_rules:
     max: 2
     default: 0
 pricing:
-  input: '0.0058'
-  output: '0.0078'
+  input: '0.0006'
+  output: '0.0006'
   unit: '0.0001'
   currency: USD

+ 5 - 5
api/core/model_runtime/model_providers/novita/llm/Nous-Hermes-2-Mixtral-8x7B-DPO.yaml

@@ -1,7 +1,7 @@
-model: Nous-Hermes-2-Mixtral-8x7B-DPO
+model: qwen/qwen-2-72b-instruct
 label:
-  zh_Hans: Nous-Hermes-2-Mixtral-8x7B-DPO
-  en_US: Nous-Hermes-2-Mixtral-8x7B-DPO
+  zh_Hans: qwen/qwen-2-72b-instruct
+  en_US: qwen/qwen-2-72b-instruct
 model_type: llm
 features:
   - agent-thought
@@ -35,7 +35,7 @@ parameter_rules:
     max: 2
     default: 0
 pricing:
-  input: '0.0027'
-  output: '0.0027'
+  input: '0.0034'
+  output: '0.0039'
   unit: '0.0001'
   currency: USD

+ 5 - 5
api/core/model_runtime/model_providers/novita/llm/llama-3.1-405b-instruct.yaml

@@ -1,7 +1,7 @@
-model: meta-llama/llama-3.1-405b-instruct
+model: qwen/qwen-2-7b-instruct
 label:
-  zh_Hans: meta-llama/llama-3.1-405b-instruct
-  en_US: meta-llama/llama-3.1-405b-instruct
+  zh_Hans: qwen/qwen-2-7b-instruct
+  en_US: qwen/qwen-2-7b-instruct
 model_type: llm
 features:
   - agent-thought
@@ -35,7 +35,7 @@ parameter_rules:
     max: 2
     default: 0
 pricing:
-  input: '0.03'
-  output: '0.05'
+  input: '0.00054'
+  output: '0.00054'
   unit: '0.0001'
   currency: USD

+ 41 - 0
api/core/model_runtime/model_providers/novita/llm/qwen-2-vl-72b-instruct.yaml

@@ -0,0 +1,41 @@
+model: qwen/qwen-2-vl-72b-instruct
+label:
+  zh_Hans: qwen/qwen-2-vl-72b-instruct
+  en_US: qwen/qwen-2-vl-72b-instruct
+model_type: llm
+features:
+  - agent-thought
+model_properties:
+  mode: chat
+  context_size: 32768
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+    min: 0
+    max: 2
+    default: 1
+  - name: top_p
+    use_template: top_p
+    min: 0
+    max: 1
+    default: 1
+  - name: max_tokens
+    use_template: max_tokens
+    min: 1
+    max: 2048
+    default: 512
+  - name: frequency_penalty
+    use_template: frequency_penalty
+    min: -2
+    max: 2
+    default: 0
+  - name: presence_penalty
+    use_template: presence_penalty
+    min: -2
+    max: 2
+    default: 0
+pricing:
+  input: '0.0045'
+  output: '0.0045'
+  unit: '0.0001'
+  currency: USD

+ 41 - 0
api/core/model_runtime/model_providers/novita/llm/qwen-2.5-72b-instruct.yaml

@@ -0,0 +1,41 @@
+model: qwen/qwen-2.5-72b-instruct
+label:
+  zh_Hans: qwen/qwen-2.5-72b-instruct
+  en_US: qwen/qwen-2.5-72b-instruct
+model_type: llm
+features:
+  - agent-thought
+model_properties:
+  mode: chat
+  context_size: 32000
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+    min: 0
+    max: 2
+    default: 1
+  - name: top_p
+    use_template: top_p
+    min: 0
+    max: 1
+    default: 1
+  - name: max_tokens
+    use_template: max_tokens
+    min: 1
+    max: 2048
+    default: 512
+  - name: frequency_penalty
+    use_template: frequency_penalty
+    min: -2
+    max: 2
+    default: 0
+  - name: presence_penalty
+    use_template: presence_penalty
+    min: -2
+    max: 2
+    default: 0
+pricing:
+  input: '0.0038'
+  output: '0.004'
+  unit: '0.0001'
+  currency: USD

+ 2 - 2
api/core/model_runtime/model_providers/novita/llm/wizardlm-2-8x22b.yaml

@@ -35,7 +35,7 @@ parameter_rules:
     max: 2
     default: 0
 pricing:
-  input: '0.0064'
-  output: '0.0064'
+  input: '0.0062'
+  output: '0.0062'
   unit: '0.0001'
   currency: USD

+ 4 - 4
api/core/model_runtime/model_providers/novita/novita.yaml

@@ -1,6 +1,6 @@
 provider: novita
 label:
-  en_US: novita.ai
+  en_US: Novita AI
 description:
   en_US: An LLM API that matches various application scenarios with high cost-effectiveness.
   zh_Hans: 适配多种海外应用场景的高性价比 LLM API
@@ -11,10 +11,10 @@ icon_large:
 background: "#eadeff"
 help:
   title:
-    en_US: Get your API key from novita.ai
-    zh_Hans: 从 novita.ai 获取 API Key
+    en_US: Get your API key from Novita AI
+    zh_Hans: 从 Novita AI 获取 API Key
   url:
-    en_US: https://novita.ai/settings#key-management?utm_source=dify&utm_medium=ch&utm_campaign=api
+    en_US: https://novita.ai/settings/key-management?utm_source=dify&utm_medium=ch&utm_campaign=api
 supported_model_types:
   - llm
 configurate_methods: