perf: yuque dataset (#5040)

* perf: yuque dataset

* doc
This commit is contained in:
Archer
2025-06-16 18:01:59 +08:00
committed by GitHub
parent 450d0a54fe
commit 7981b61ca9
9 changed files with 139 additions and 53 deletions

View File

@@ -1,6 +1,72 @@
{
"provider": "Doubao",
"list": [
{
"model": "Doubao-Seed-1.6",
"name": "Doubao-Seed-1.6",
"maxContext": 220000,
"maxResponse": 16000,
"quoteMaxToken": 220000,
"maxTemperature": 1,
"showTopP": true,
"showStopSign": true,
"vision": true,
"toolChoice": true,
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInQueryExtension": true,
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {},
"type": "llm"
},
{
"model": "Doubao-Seed-1.6-thinking",
"name": "Doubao-Seed-1.6-thinking",
"maxContext": 220000,
"maxResponse": 16000,
"quoteMaxToken": 220000,
"maxTemperature": 1,
"showTopP": true,
"showStopSign": true,
"vision": true,
"toolChoice": true,
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInQueryExtension": true,
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {},
"type": "llm"
},
{
"model": "Doubao-Seed-1.6-flash",
"name": "Doubao-Seed-1.6-flash",
"maxContext": 220000,
"maxResponse": 16000,
"quoteMaxToken": 220000,
"maxTemperature": 1,
"showTopP": true,
"showStopSign": true,
"vision": true,
"toolChoice": true,
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInQueryExtension": true,
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {},
"type": "llm"
},
{
"model": "Doubao-1.5-lite-32k",
"name": "Doubao-1.5-lite-32k",

View File

@@ -4,9 +4,9 @@
{
"model": "qwen-max",
"name": "Qwen-max",
"maxContext": 32000,
"maxResponse": 4000,
"quoteMaxToken": 6000,
"maxContext": 128000,
"maxResponse": 8000,
"quoteMaxToken": 120000,
"maxTemperature": 1,
"vision": false,
"toolChoice": true,
@@ -27,10 +27,10 @@
{
"model": "qwen-vl-max",
"name": "qwen-vl-max",
"maxContext": 32000,
"maxResponse": 2000,
"quoteMaxToken": 20000,
"maxTemperature": 1.2,
"maxContext": 128000,
"maxResponse": 8000,
"quoteMaxToken": 120000,
"maxTemperature": 1,
"vision": true,
"toolChoice": false,
"functionCall": false,
@@ -49,9 +49,9 @@
{
"model": "qwen-plus",
"name": "Qwen-plus",
"maxContext": 64000,
"maxContext": 128000,
"maxResponse": 8000,
"quoteMaxToken": 60000,
"quoteMaxToken": 120000,
"maxTemperature": 1,
"vision": false,
"toolChoice": true,
@@ -72,10 +72,10 @@
{
"model": "qwen-vl-plus",
"name": "qwen-vl-plus",
"maxContext": 32000,
"maxResponse": 2000,
"quoteMaxToken": 20000,
"maxTemperature": 1.2,
"maxContext": 128000,
"maxResponse": 8000,
"quoteMaxToken": 120000,
"maxTemperature": 1,
"vision": true,
"toolChoice": false,
"functionCall": false,
@@ -92,9 +92,9 @@
{
"model": "qwen-turbo",
"name": "Qwen-turbo",
"maxContext": 128000,
"maxContext": 1000000,
"maxResponse": 8000,
"quoteMaxToken": 100000,
"quoteMaxToken": 1000000,
"maxTemperature": 1,
"vision": false,
"toolChoice": true,
@@ -487,9 +487,9 @@
{
"model": "qwen-long",
"name": "qwen-long",
"maxContext": 100000,
"maxContext": 10000000,
"maxResponse": 6000,
"quoteMaxToken": 10000,
"quoteMaxToken": 10000000,
"maxTemperature": 1,
"vision": false,
"toolChoice": false,

View File

@@ -106,7 +106,7 @@ export const useApiDatasetRequest = ({ apiServer }: { apiServer: APIFileServer }
const formattedFiles = files.map((file) => ({
...file,
hasChild: file.type === 'folder'
hasChild: file.hasChild ?? file.type === 'folder'
}));
return formattedFiles;

View File

@@ -198,6 +198,7 @@ export const useYuqueDatasetRequest = ({ yuqueServer }: { yuqueServer: YuqueServ
}: {
apiFileId: string;
}): Promise<ApiFileReadContentResponse> => {
if (typeof apiFileId !== 'string') return Promise.reject('Invalid file id');
const [parentId, fileId] = apiFileId.split(/-(.*?)-(.*)/);
const data = await request<{ title: string; body: string }>(

View File

@@ -167,7 +167,7 @@ export const readApiServerFileContent = async ({
};
export const rawText2Chunks = async ({
rawText,
rawText = '',
chunkTriggerType = ChunkTriggerConfigTypeEnum.minSize,
chunkTriggerMinSize = 1000,
backupParse,