From ec8e2512bc9a0c6e312fb151063e96246e5ccc33 Mon Sep 17 00:00:00 2001 From: Robin Wang Date: Fri, 2 Feb 2024 12:20:16 +0800 Subject: [PATCH] change gpt-4-turbo maxResponse configuration template (#814) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit 用最新的配置文件4.6.8 ,对话选gpt-4-turbo 报错: null max_tokens is too large: 62500. This model supports at most 4096 completion tokens, whereas you provided 62500. (request id: 20240202110253407344738SmDnkwX1) 原因是官方gpt-4-turbo 最大的返回token 4096. --- docSite/content/docs/development/configuration.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docSite/content/docs/development/configuration.md b/docSite/content/docs/development/configuration.md index bf1916933..93bc62579 100644 --- a/docSite/content/docs/development/configuration.md +++ b/docSite/content/docs/development/configuration.md @@ -224,7 +224,7 @@ llm模型全部合并 "model": "gpt-4-0125-preview", "name": "gpt-4-turbo", "maxContext": 125000, - "maxResponse": 125000, + "maxResponse": 4000, "quoteMaxToken": 100000, "maxTemperature": 1.2, "inputPrice": 0,