From 7a9cf4ce9eae62b21805528e5f1959597988f7b1 Mon Sep 17 00:00:00 2001 From: Archer <545436317@qq.com> Date: Tue, 15 Apr 2025 10:49:46 +0800 Subject: [PATCH] add system model (#4540) --- deploy/docker/docker-compose-milvus.yml | 8 +- deploy/docker/docker-compose-pgvector.yml | 8 +- deploy/docker/docker-compose-zilliz.yml | 8 +- .../zh-cn/docs/development/upgrading/495.md | 12 ++- .../core/ai/config/provider/Gemini.json | 24 ++++++ .../service/core/ai/config/provider/Grok.json | 74 +++++++++++++++- .../core/ai/config/provider/OpenAI.json | 86 ++++++++++++++++--- 7 files changed, 195 insertions(+), 25 deletions(-) diff --git a/deploy/docker/docker-compose-milvus.yml b/deploy/docker/docker-compose-milvus.yml index 8877f7763..f2dfdaeb5 100644 --- a/deploy/docker/docker-compose-milvus.yml +++ b/deploy/docker/docker-compose-milvus.yml @@ -126,15 +126,15 @@ services: # fastgpt sandbox: container_name: sandbox - image: ghcr.io/labring/fastgpt-sandbox:v4.9.4 # git - # image: registry.cn-hangzhou.aliyuncs.com/fastgpt/fastgpt-sandbox:v4.9.4 # 阿里云 + image: ghcr.io/labring/fastgpt-sandbox:v4.9.5 # git + # image: registry.cn-hangzhou.aliyuncs.com/fastgpt/fastgpt-sandbox:v4.9.5 # 阿里云 networks: - fastgpt restart: always fastgpt: container_name: fastgpt - image: ghcr.io/labring/fastgpt:v4.9.4 # git - # image: registry.cn-hangzhou.aliyuncs.com/fastgpt/fastgpt:v4.9.4 # 阿里云 + image: ghcr.io/labring/fastgpt:v4.9.5 # git + # image: registry.cn-hangzhou.aliyuncs.com/fastgpt/fastgpt:v4.9.5 # 阿里云 ports: - 3000:3000 networks: diff --git a/deploy/docker/docker-compose-pgvector.yml b/deploy/docker/docker-compose-pgvector.yml index 05ebdf936..3f0a14921 100644 --- a/deploy/docker/docker-compose-pgvector.yml +++ b/deploy/docker/docker-compose-pgvector.yml @@ -85,15 +85,15 @@ services: # fastgpt sandbox: container_name: sandbox - image: ghcr.io/labring/fastgpt-sandbox:v4.9.4 # git - # image: registry.cn-hangzhou.aliyuncs.com/fastgpt/fastgpt-sandbox:v4.9.4 # 阿里云 + image: ghcr.io/labring/fastgpt-sandbox:v4.9.5 # git + # image: registry.cn-hangzhou.aliyuncs.com/fastgpt/fastgpt-sandbox:v4.9.5 # 阿里云 networks: - fastgpt restart: always fastgpt: container_name: fastgpt - image: ghcr.io/labring/fastgpt:v4.9.4 # git - # image: registry.cn-hangzhou.aliyuncs.com/fastgpt/fastgpt:v4.9.4 # 阿里云 + image: ghcr.io/labring/fastgpt:v4.9.5 # git + # image: registry.cn-hangzhou.aliyuncs.com/fastgpt/fastgpt:v4.9.5 # 阿里云 ports: - 3000:3000 networks: diff --git a/deploy/docker/docker-compose-zilliz.yml b/deploy/docker/docker-compose-zilliz.yml index da80b4cb8..777d789cf 100644 --- a/deploy/docker/docker-compose-zilliz.yml +++ b/deploy/docker/docker-compose-zilliz.yml @@ -66,15 +66,15 @@ services: sandbox: container_name: sandbox - image: ghcr.io/labring/fastgpt-sandbox:v4.9.4 # git - # image: registry.cn-hangzhou.aliyuncs.com/fastgpt/fastgpt-sandbox:v4.9.4 # 阿里云 + image: ghcr.io/labring/fastgpt-sandbox:v4.9.5 # git + # image: registry.cn-hangzhou.aliyuncs.com/fastgpt/fastgpt-sandbox:v4.9.5 # 阿里云 networks: - fastgpt restart: always fastgpt: container_name: fastgpt - image: ghcr.io/labring/fastgpt:v4.9.4 # git - # image: registry.cn-hangzhou.aliyuncs.com/fastgpt/fastgpt:v4.9.4 # 阿里云 + image: ghcr.io/labring/fastgpt:v4.9.5 # git + # image: registry.cn-hangzhou.aliyuncs.com/fastgpt/fastgpt:v4.9.5 # 阿里云 ports: - 3000:3000 networks: diff --git a/docSite/content/zh-cn/docs/development/upgrading/495.md b/docSite/content/zh-cn/docs/development/upgrading/495.md index 9d80440f9..a8f1f5c70 100644 --- a/docSite/content/zh-cn/docs/development/upgrading/495.md +++ b/docSite/content/zh-cn/docs/development/upgrading/495.md @@ -1,5 +1,5 @@ --- -title: 'V4.9.5(进行中)' +title: 'V4.9.5' description: 'FastGPT V4.9.5 更新说明' icon: 'upgrade' draft: false @@ -7,6 +7,15 @@ toc: true weight: 795 --- +## 升级指南 +### 1. 做好数据备份 + +### 2. 更新镜像 tag + +- 更新 FastGPT 镜像 tag: v4.9.5 +- 更新 FastGPT 商业版镜像 tag: v4.9.5 +- Sandbox 无需更新 +- AIProxy 无需更新 ## 🚀 新增内容 @@ -18,6 +27,7 @@ weight: 795 ## ⚙️ 优化 1. 繁体中文翻译。 +2. Arm 镜像打包 ## 🐛 修复 diff --git a/packages/service/core/ai/config/provider/Gemini.json b/packages/service/core/ai/config/provider/Gemini.json index aef0a9986..1df074c74 100644 --- a/packages/service/core/ai/config/provider/Gemini.json +++ b/packages/service/core/ai/config/provider/Gemini.json @@ -1,6 +1,30 @@ { "provider": "Gemini", "list": [ + { + "model": "gemini-2.5-pro-exp-03-25", + "name": "gemini-2.5-pro-exp-03-25", + "maxContext": 1000000, + "maxResponse": 63000, + "quoteMaxToken": 1000000, + "maxTemperature": 1, + "vision": true, + "toolChoice": true, + "functionCall": false, + "defaultSystemChatPrompt": "", + "datasetProcess": true, + "usedInClassify": true, + "customCQPrompt": "", + "usedInExtractFields": true, + "usedInQueryExtension": true, + "customExtractPrompt": "", + "usedInToolCall": true, + "defaultConfig": {}, + "fieldMap": {}, + "type": "llm", + "showTopP": true, + "showStopSign": true + }, { "model": "gemini-2.0-flash", "name": "gemini-2.0-flash", diff --git a/packages/service/core/ai/config/provider/Grok.json b/packages/service/core/ai/config/provider/Grok.json index e60bab604..197bb3e0a 100644 --- a/packages/service/core/ai/config/provider/Grok.json +++ b/packages/service/core/ai/config/provider/Grok.json @@ -1,6 +1,54 @@ { "provider": "Grok", "list": [ + { + "model": "grok-3-mini", + "name": "grok-3-mini", + "maxContext": 128000, + "maxResponse": 8000, + "quoteMaxToken": 128000, + "maxTemperature": 1, + "showTopP": true, + "showStopSign": true, + "vision": false, + "toolChoice": true, + "functionCall": false, + "defaultSystemChatPrompt": "", + "datasetProcess": true, + "usedInClassify": true, + "customCQPrompt": "", + "usedInExtractFields": true, + "usedInQueryExtension": true, + "customExtractPrompt": "", + "usedInToolCall": true, + "defaultConfig": {}, + "fieldMap": {}, + "type": "llm" + }, + { + "model": "grok-3-mini-fast", + "name": "grok-3-mini-fast", + "maxContext": 128000, + "maxResponse": 8000, + "quoteMaxToken": 128000, + "maxTemperature": 1, + "showTopP": true, + "showStopSign": true, + "vision": false, + "toolChoice": true, + "functionCall": false, + "defaultSystemChatPrompt": "", + "datasetProcess": true, + "usedInClassify": true, + "customCQPrompt": "", + "usedInExtractFields": true, + "usedInQueryExtension": true, + "customExtractPrompt": "", + "usedInToolCall": true, + "defaultConfig": {}, + "fieldMap": {}, + "type": "llm" + }, { "model": "grok-3", "name": "grok-3", @@ -11,7 +59,31 @@ "showTopP": true, "showStopSign": true, "vision": false, - "toolChoice": false, + "toolChoice": true, + "functionCall": false, + "defaultSystemChatPrompt": "", + "datasetProcess": true, + "usedInClassify": true, + "customCQPrompt": "", + "usedInExtractFields": true, + "usedInQueryExtension": true, + "customExtractPrompt": "", + "usedInToolCall": true, + "defaultConfig": {}, + "fieldMap": {}, + "type": "llm" + }, + { + "model": "grok-3-fast", + "name": "grok-3-fast", + "maxContext": 128000, + "maxResponse": 8000, + "quoteMaxToken": 128000, + "maxTemperature": 1, + "showTopP": true, + "showStopSign": true, + "vision": false, + "toolChoice": true, "functionCall": false, "defaultSystemChatPrompt": "", "datasetProcess": true, diff --git a/packages/service/core/ai/config/provider/OpenAI.json b/packages/service/core/ai/config/provider/OpenAI.json index 324958f5c..d086d717c 100644 --- a/packages/service/core/ai/config/provider/OpenAI.json +++ b/packages/service/core/ai/config/provider/OpenAI.json @@ -1,6 +1,78 @@ { "provider": "OpenAI", "list": [ + { + "model": "gpt-4.1", + "name": "gpt-4.1", + "maxContext": 1000000, + "maxResponse": 32000, + "quoteMaxToken": 1000000, + "maxTemperature": 1.2, + "showTopP": true, + "responseFormatList": ["text", "json_object", "json_schema"], + "showStopSign": true, + "vision": true, + "toolChoice": true, + "functionCall": true, + "defaultSystemChatPrompt": "", + "datasetProcess": true, + "usedInClassify": true, + "customCQPrompt": "", + "usedInExtractFields": true, + "customExtractPrompt": "", + "usedInToolCall": true, + "defaultConfig": {}, + "fieldMap": {}, + "type": "llm" + }, + { + "model": "gpt-4.1-mini", + "name": "gpt-4.1-mini", + "maxContext": 1000000, + "maxResponse": 32000, + "quoteMaxToken": 1000000, + "maxTemperature": 1.2, + "showTopP": true, + "responseFormatList": ["text", "json_object", "json_schema"], + "showStopSign": true, + "vision": true, + "toolChoice": true, + "functionCall": true, + "defaultSystemChatPrompt": "", + "datasetProcess": true, + "usedInClassify": true, + "customCQPrompt": "", + "usedInExtractFields": true, + "customExtractPrompt": "", + "usedInToolCall": true, + "defaultConfig": {}, + "fieldMap": {}, + "type": "llm" + }, + { + "model": "gpt-4.1-nano", + "name": "gpt-4.1-nano", + "maxContext": 1000000, + "maxResponse": 32000, + "quoteMaxToken": 1000000, + "maxTemperature": 1.2, + "showTopP": true, + "responseFormatList": ["text", "json_object", "json_schema"], + "showStopSign": true, + "vision": true, + "toolChoice": true, + "functionCall": true, + "defaultSystemChatPrompt": "", + "datasetProcess": true, + "usedInClassify": true, + "customCQPrompt": "", + "usedInExtractFields": true, + "customExtractPrompt": "", + "usedInToolCall": true, + "defaultConfig": {}, + "fieldMap": {}, + "type": "llm" + }, { "model": "gpt-4o-mini", "name": "GPT-4o-mini", @@ -9,11 +81,7 @@ "quoteMaxToken": 60000, "maxTemperature": 1.2, "showTopP": true, - "responseFormatList": [ - "text", - "json_object", - "json_schema" - ], + "responseFormatList": ["text", "json_object", "json_schema"], "showStopSign": true, "vision": true, "toolChoice": true, @@ -37,11 +105,7 @@ "quoteMaxToken": 60000, "maxTemperature": 1.2, "showTopP": true, - "responseFormatList": [ - "text", - "json_object", - "json_schema" - ], + "responseFormatList": ["text", "json_object", "json_schema"], "showStopSign": true, "vision": true, "toolChoice": true, @@ -275,4 +339,4 @@ "type": "stt" } ] -} \ No newline at end of file +}