Browse Source

feat: add support for Vertex AI Gemini 1.5 002 and experimental models (#8767)

Pan, Wen-Ming 6 months ago
parent
commit
02ff6cca70

+ 1 - 1
api/core/model_runtime/model_providers/vertex_ai/llm/gemini-1.5-flash.yaml

@@ -1,6 +1,6 @@
 model: gemini-1.5-flash-001
 label:
-  en_US: Gemini 1.5 Flash
+  en_US: Gemini 1.5 Flash 001
 model_type: llm
 features:
   - agent-thought

+ 37 - 0
api/core/model_runtime/model_providers/vertex_ai/llm/gemini-1.5-flash-002.yaml

@@ -0,0 +1,37 @@
+model: gemini-1.5-flash-002
+label:
+  en_US: Gemini 1.5 Flash 002
+model_type: llm
+features:
+  - agent-thought
+  - vision
+model_properties:
+  mode: chat
+  context_size: 1048576
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+  - name: top_p
+    use_template: top_p
+  - name: top_k
+    label:
+      en_US: Top k
+    type: int
+    help:
+      en_US: Only sample from the top K options for each subsequent token.
+    required: false
+  - name: presence_penalty
+    use_template: presence_penalty
+  - name: frequency_penalty
+    use_template: frequency_penalty
+  - name: max_output_tokens
+    use_template: max_tokens
+    required: true
+    default: 8192
+    min: 1
+    max: 8192
+pricing:
+  input: '0.00'
+  output: '0.00'
+  unit: '0.000001'
+  currency: USD

+ 1 - 1
api/core/model_runtime/model_providers/vertex_ai/llm/gemini-1.5-pro.yaml

@@ -1,6 +1,6 @@
 model: gemini-1.5-pro-001
 label:
-  en_US: Gemini 1.5 Pro
+  en_US: Gemini 1.5 Pro 001
 model_type: llm
 features:
   - agent-thought

+ 37 - 0
api/core/model_runtime/model_providers/vertex_ai/llm/gemini-1.5-pro-002.yaml

@@ -0,0 +1,37 @@
+model: gemini-1.5-pro-002
+label:
+  en_US: Gemini 1.5 Pro 002
+model_type: llm
+features:
+  - agent-thought
+  - vision
+model_properties:
+  mode: chat
+  context_size: 1048576
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+  - name: top_p
+    use_template: top_p
+  - name: top_k
+    label:
+      en_US: Top k
+    type: int
+    help:
+      en_US: Only sample from the top K options for each subsequent token.
+    required: false
+  - name: presence_penalty
+    use_template: presence_penalty
+  - name: frequency_penalty
+    use_template: frequency_penalty
+  - name: max_output_tokens
+    use_template: max_tokens
+    required: true
+    default: 8192
+    min: 1
+    max: 8192
+pricing:
+  input: '0.00'
+  output: '0.00'
+  unit: '0.000001'
+  currency: USD

+ 37 - 0
api/core/model_runtime/model_providers/vertex_ai/llm/gemini-flash-experimental.yaml

@@ -0,0 +1,37 @@
+model: gemini-flash-experimental
+label:
+  en_US: Gemini Flash Experimental
+model_type: llm
+features:
+  - agent-thought
+  - vision
+model_properties:
+  mode: chat
+  context_size: 1048576
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+  - name: top_p
+    use_template: top_p
+  - name: top_k
+    label:
+      en_US: Top k
+    type: int
+    help:
+      en_US: Only sample from the top K options for each subsequent token.
+    required: false
+  - name: presence_penalty
+    use_template: presence_penalty
+  - name: frequency_penalty
+    use_template: frequency_penalty
+  - name: max_output_tokens
+    use_template: max_tokens
+    required: true
+    default: 8192
+    min: 1
+    max: 8192
+pricing:
+  input: '0.00'
+  output: '0.00'
+  unit: '0.000001'
+  currency: USD

+ 37 - 0
api/core/model_runtime/model_providers/vertex_ai/llm/gemini-pro-experimental.yaml

@@ -0,0 +1,37 @@
+model: gemini-pro-experimental
+label:
+  en_US: Gemini Pro Experimental
+model_type: llm
+features:
+  - agent-thought
+  - vision
+model_properties:
+  mode: chat
+  context_size: 1048576
+parameter_rules:
+  - name: temperature
+    use_template: temperature
+  - name: top_p
+    use_template: top_p
+  - name: top_k
+    label:
+      en_US: Top k
+    type: int
+    help:
+      en_US: Only sample from the top K options for each subsequent token.
+    required: false
+  - name: presence_penalty
+    use_template: presence_penalty
+  - name: frequency_penalty
+    use_template: frequency_penalty
+  - name: max_output_tokens
+    use_template: max_tokens
+    required: true
+    default: 8192
+    min: 1
+    max: 8192
+pricing:
+  input: '0.00'
+  output: '0.00'
+  unit: '0.000001'
+  currency: USD