mirror of
https://github.com/labring/FastGPT.git
synced 2026-05-07 01:02:55 +08:00
76d6234de6
* Agent features (#6345) * Test agent (#6220) * squash: compress all commits into one * feat: plan response in ui * response ui * perf: agent config * merge * tool select ux * perf: chat ui * perf: agent editform * tmp code * feat: save chat * Complete agent parent (#6049) * add role and tools filling * add: file-upload --------- Co-authored-by: xxyyh <2289112474@qq> * perf: top agent code * top agent (#6062) Co-authored-by: xxyyh <2289112474@qq> * fix: ts * skill editor ui * ui * perf: rewrite type with zod * skill edit ui * skill agent (#6089) * cp skill chat * rebasefdf933dand add skill chat * 1. skill 的 CRUD 2. skill 的信息渲染到前端界面 * solve comment * remove chatid and chatItemId * skill match * perf: skill manage * fix: ts --------- Co-authored-by: xxyyh <2289112474@qq> Co-authored-by: archer <545436317@qq.com> * fix: ts * fix: loop import * skill tool config (#6114) Co-authored-by: xxyyh <2289112474@qq> * feat: load tool in agent * skill memory (#6126) Co-authored-by: xxyyh <2289112474@qq> * perf: agent skill editor * perf: helperbot ui * agent code * perf: context * fix: request context * agent usage * perf: agent context and pause * perf: plan response * Test agent sigle skill (#6184) * feat:top box fill * prompt fix --------- Co-authored-by: xxyyh <2289112474@qq> * perf: agent chat ui * Test agent new (#6219) * have-replan * agent --------- Co-authored-by: xxyyh <2289112474@qq> * fix: ts --------- Co-authored-by: YeYuheng <57035043+YYH211@users.noreply.github.com> Co-authored-by: xxyyh <2289112474@qq> * feat: consolidate agent and MCP improvements This commit consolidates 17 commits including: - MCP tools enhancements and fixes - Agent system improvements and optimizations - Auth limit and prompt updates - Tool response compression and error tracking - Simple app adaptation - Code quality improvements (TypeScript, ESLint, Zod) - Version type migration to schema - Remove deprecated useRequest2 - Add LLM error tracking - Toolset ID validation fixes --------- Co-authored-by: YeYuheng <57035043+YYH211@users.noreply.github.com> Co-authored-by: xxyyh <2289112474@qq> * fix: transform avatar copy;perf: filter invalid tool * update llm response storage time * fix: openapi schema * update skill desc * feat: cache hit data * i18n * lock * chat logs support error filter & user search (#6373) * chat log support searching by user name * support error filter * fix * fix overflow * optimize * fix init script * fix * perf: get log users * updat ecomment * fix: ts * fix: test --------- Co-authored-by: archer <545436317@qq.com> * Fix: agent (#6376) * Agent features (#6345) * Test agent (#6220) * squash: compress all commits into one * feat: plan response in ui * response ui * perf: agent config * merge * tool select ux * perf: chat ui * perf: agent editform * tmp code * feat: save chat * Complete agent parent (#6049) * add role and tools filling * add: file-upload --------- Co-authored-by: xxyyh <2289112474@qq> * perf: top agent code * top agent (#6062) Co-authored-by: xxyyh <2289112474@qq> * fix: ts * skill editor ui * ui * perf: rewrite type with zod * skill edit ui * skill agent (#6089) * cp skill chat * rebasefdf933dand add skill chat * 1. skill 的 CRUD 2. skill 的信息渲染到前端界面 * solve comment * remove chatid and chatItemId * skill match * perf: skill manage * fix: ts --------- Co-authored-by: xxyyh <2289112474@qq> Co-authored-by: archer <545436317@qq.com> * fix: ts * fix: loop import * skill tool config (#6114) Co-authored-by: xxyyh <2289112474@qq> * feat: load tool in agent * skill memory (#6126) Co-authored-by: xxyyh <2289112474@qq> * perf: agent skill editor * perf: helperbot ui * agent code * perf: context * fix: request context * agent usage * perf: agent context and pause * perf: plan response * Test agent sigle skill (#6184) * feat:top box fill * prompt fix --------- Co-authored-by: xxyyh <2289112474@qq> * perf: agent chat ui * Test agent new (#6219) * have-replan * agent --------- Co-authored-by: xxyyh <2289112474@qq> * fix: ts --------- Co-authored-by: YeYuheng <57035043+YYH211@users.noreply.github.com> Co-authored-by: xxyyh <2289112474@qq> * feat: consolidate agent and MCP improvements This commit consolidates 17 commits including: - MCP tools enhancements and fixes - Agent system improvements and optimizations - Auth limit and prompt updates - Tool response compression and error tracking - Simple app adaptation - Code quality improvements (TypeScript, ESLint, Zod) - Version type migration to schema - Remove deprecated useRequest2 - Add LLM error tracking - Toolset ID validation fixes --------- Co-authored-by: YeYuheng <57035043+YYH211@users.noreply.github.com> Co-authored-by: xxyyh <2289112474@qq> * 1. 把辅助生成前端上的 system prompt 加入到上下文中 2. mcp工具的前端渲染(图标) 3. 文件读取工具和文件上传进行关联 4. 添加了辅助生成返回格式出错的重试方案 5. ask 不出现在 plan 步骤中 6. 添加了辅助生成的头像和交互 UI * fix:read_file * helperbot ui * ts error * helper ui * delete Unused import * perf: helper bot * lock --------- Co-authored-by: Archer <545436317@qq.com> Co-authored-by: xxyyh <2289112474@qq> * fix date variable required & model auth (#6386) * fix date variable required & model auth * doc * feat: add chat id to finish callback * fix: iphone safari shareId (#6387) * fix: iphone safari shareId * fix: mcp file list can't setting * fix: reason output field * fix: skip JSON validation for HTTP tool body with variable (#6392) * fix: skip JSON validation for HTTP tool body with variable * doc * workflow fitview * perf: selecting memory * perf: cp api * ui * perf: toolcall auto adapt * fix: catch workflow error * fix: ts * perf: pagination type * remove * ignore * update doc * fix: simple app tool select * add default avatar to logs user * perf: loading user * select dataset ui * rename version * feat: add global/common test * perf: packages/global/common test * feat: package/global/ai,app test * add global/chat test * global/core test * global/core test * feat: packages/global all test * perf: test * add server api test * perf: init shell * perf: init4150 shell * remove invalid code * update doc * remove log * fix: chat effect * fix: plan fake tool (#6398) * 1. 提示词防注入功能 2. 无工具不进入 plan,防止虚拟工具生成 * Agent-dataset * dataset * dataset presetInfo * prefix * perf: prompt --------- Co-authored-by: xxyyh <2289112474@qq> Co-authored-by: archer <545436317@qq.com> * fix: review * adapt kimi2.5 think toolcall * feat: invoke fastgpt user info (#6403) feat: invoke fastgpt user info * fix: invoke fastgpt user info return orgs (#6404) * skill and version * retry helperbot (#6405) Co-authored-by: xxyyh <2289112474@qq> * update template * remove log * doc * update doc * doc * perf: internal ip check * adapt get paginationRecords * tool call adapt * fix: test * doc * fix: agent initial version * adapt completions v1 * feat: instrumentation check * rename skill * add workflow demo mode tracks (#6407) * chore: 统一 skills 目录命名为小写 将 .claude/Skills/ 重命名为 .claude/skills/ 以保持命名一致性。 Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com> * add workflow demo mode tracks * code * optimize * fix: improve workflowDemoTrack based on PR review - Add comment to empty catch block for maintainability - Add @param docs to onDemoChange clarifying nodeCount usage - Replace silent .catch with console.debug for dev debugging - Handle appId changes by reporting old data before re-init Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com> --------- Co-authored-by: archer <545436317@qq.com> Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com> * remove repeat skill * fix(workflow): filter out orphan edges to prevent runtime errors (#6399) * fix(workflow): filter out orphan edges to prevent runtime errors Runtime edges that reference non-existent nodes (orphan edges) can cause unexpected behavior or crashes during workflow dispatch. This change adds a pre-check to filter out such edges before execution begins, ensuring system stability even with inconsistent graph data. * fix(workflow): enhance orphan edge filtering with logging and tests - Refactor: Extract logic to 'filterOrphanEdges' in utils.ts for better reusability - Feat: Add performance monitoring (warn if >100ms) and comprehensive logging - Feat: Support detailed edge inspection in debug mode - Docs: Add JSDoc explaining causes of orphan edges (migration, manual edits) - Test: Add unit tests covering edge cases and performance (1000 edges) Addresses PR review feedback regarding logging, variable naming, and testing." * move code * move code * add more unit test --------- Co-authored-by: archer <545436317@qq.com> * test * perf: test * add server/common/string test * fix: resolve $ref references in MCP tool input schemas (#6395) (#6409) * fix: resolve $ref references in MCP tool input schemas (#6395) * add test code --------- Co-authored-by: archer <545436317@qq.com> * chore(docs): add fastgpt, fastgpt-plugin version choice guide (#6411) * chore(doc): add fastgpt version description * doc * doc --------- Co-authored-by: archer <545436317@qq.com> * fix:dataset cite and description info (#6410) * 1. 添加知识库引用(plan 步骤和直接知识库调用) 2. 提示词框中的@知识库工具 3. plan 中 step 的 description dataset_search 改为中文 * fix: i18n * prompt * prompt --------- Co-authored-by: xxyyh <2289112474@qq> * fix: tool call * perf: workflow props * fix: merge ECharts toolbox options instead of overwriting (#6269) (#6412) * feat: integrate logtape and otel (#6400) * fix: deps * feat(logger): integrate logtape and otel * wip(log): add basic infras logs * wip(log): add request id and inject it into context * wip(log): add basic tx logs * wip(log): migrate * wip(log): category * wip(log): more sub category * fix: type * fix: sessionRun * fix: export getLogger from client.ts * chore: improve logs * docs: update signoz and changelog * change type * fix: ts * remove skill.md * fix: lockfile specifier * fix: test --------- Co-authored-by: archer <545436317@qq.com> * init log * doc * remove invalid log * fix: review * template * replace new log * fix: ts * remove log * chore: migrate all addLog to logtape * move skill * chore: migrate all addLog to logtape (#6417) * update skill * remove log * fix: tool check --------- Co-authored-by: YeYuheng <57035043+YYH211@users.noreply.github.com> Co-authored-by: xxyyh <2289112474@qq> Co-authored-by: heheer <heheer@sealos.io> Co-authored-by: Finley Ge <32237950+FinleyGe@users.noreply.github.com> Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com> Co-authored-by: xuyafei1996 <54217479+xuyafei1996@users.noreply.github.com> Co-authored-by: ToukoYui <2331631097@qq.com> Co-authored-by: roy <whoeverimf5@gmail.com>
300 lines
10 KiB
TypeScript
300 lines
10 KiB
TypeScript
import type { SystemDefaultModelType, SystemModelItemType } from '../type';
|
|
import { ModelTypeEnum } from '@fastgpt/global/core/ai/constants';
|
|
import { MongoSystemModel } from './schema';
|
|
import {
|
|
type LLMModelItemType,
|
|
type EmbeddingModelItemType,
|
|
type TTSModelType,
|
|
type STTModelType,
|
|
type RerankModelItemType
|
|
} from '@fastgpt/global/core/ai/model.schema';
|
|
import { debounce } from 'lodash';
|
|
import { getModelProvider } from '../../../core/app/provider/controller';
|
|
import { findModelFromAlldata } from '../model';
|
|
import {
|
|
reloadFastGPTConfigBuffer,
|
|
updateFastGPTConfigBuffer
|
|
} from '../../../common/system/config/controller';
|
|
import { delay } from '@fastgpt/global/common/system/utils';
|
|
import { pluginClient } from '../../../thirdProvider/fastgptPlugin';
|
|
import { setCron } from '../../../common/system/cron';
|
|
import { preloadModelProviders } from '../../../core/app/provider/controller';
|
|
import { refreshVersionKey } from '../../../common/cache';
|
|
import { SystemCacheKeyEnum } from '../../../common/cache/type';
|
|
import { getLogger, LogCategories } from '../../../common/logger';
|
|
|
|
export const loadSystemModels = async (init = false, language = 'en') => {
|
|
if (!init && global.systemModelList) return;
|
|
|
|
try {
|
|
await preloadModelProviders();
|
|
} catch (error) {
|
|
const logger = getLogger(LogCategories.MODULE.AI.CONFIG);
|
|
logger.error('System model provider preload failed', { error });
|
|
return Promise.reject(error);
|
|
}
|
|
|
|
let _systemModelList: SystemModelItemType[] = [];
|
|
let _systemActiveModelList: SystemModelItemType[] = [];
|
|
let _llmModelMap = new Map<string, LLMModelItemType>();
|
|
let _embeddingModelMap = new Map<string, EmbeddingModelItemType>();
|
|
let _ttsModelMap = new Map<string, TTSModelType>();
|
|
let _sttModelMap = new Map<string, STTModelType>();
|
|
let _reRankModelMap = new Map<string, RerankModelItemType>();
|
|
let _systemDefaultModel: SystemDefaultModelType = {};
|
|
|
|
if (!global.systemModelList) {
|
|
global.systemModelList = [];
|
|
global.systemActiveModelList = [];
|
|
global.llmModelMap = new Map<string, LLMModelItemType>();
|
|
global.embeddingModelMap = new Map<string, EmbeddingModelItemType>();
|
|
global.ttsModelMap = new Map<string, TTSModelType>();
|
|
global.sttModelMap = new Map<string, STTModelType>();
|
|
global.reRankModelMap = new Map<string, RerankModelItemType>();
|
|
global.systemDefaultModel = {};
|
|
global.systemActiveDesensitizedModels = [];
|
|
}
|
|
|
|
const pushModel = (model: SystemModelItemType) => {
|
|
_systemModelList.push(model);
|
|
|
|
// Add default value
|
|
if (model.type === ModelTypeEnum.llm) {
|
|
model.datasetProcess = model.datasetProcess ?? true;
|
|
model.usedInClassify = model.usedInClassify ?? true;
|
|
model.usedInExtractFields = model.usedInExtractFields ?? true;
|
|
model.usedInToolCall = model.usedInToolCall ?? true;
|
|
model.useInEvaluation = model.useInEvaluation ?? true;
|
|
}
|
|
|
|
if (model.isActive) {
|
|
_systemActiveModelList.push(model);
|
|
|
|
if (model.type === ModelTypeEnum.llm) {
|
|
_llmModelMap.set(model.model, model);
|
|
_llmModelMap.set(model.name, model);
|
|
if (model.isDefault) {
|
|
_systemDefaultModel.llm = model;
|
|
}
|
|
if (model.isDefaultDatasetTextModel) {
|
|
_systemDefaultModel.datasetTextLLM = model;
|
|
}
|
|
if (model.isDefaultDatasetImageModel) {
|
|
_systemDefaultModel.datasetImageLLM = model;
|
|
}
|
|
if (model.model === process.env.HELPER_BOT_MODEL) {
|
|
_systemDefaultModel.helperBotLLM = model;
|
|
}
|
|
} else if (model.type === ModelTypeEnum.embedding) {
|
|
_embeddingModelMap.set(model.model, model);
|
|
_embeddingModelMap.set(model.name, model);
|
|
if (model.isDefault) {
|
|
_systemDefaultModel.embedding = model;
|
|
}
|
|
} else if (model.type === ModelTypeEnum.tts) {
|
|
_ttsModelMap.set(model.model, model);
|
|
_ttsModelMap.set(model.name, model);
|
|
if (model.isDefault) {
|
|
_systemDefaultModel.tts = model;
|
|
}
|
|
} else if (model.type === ModelTypeEnum.stt) {
|
|
_sttModelMap.set(model.model, model);
|
|
_sttModelMap.set(model.name, model);
|
|
if (model.isDefault) {
|
|
_systemDefaultModel.stt = model;
|
|
}
|
|
} else if (model.type === ModelTypeEnum.rerank) {
|
|
_reRankModelMap.set(model.model, model);
|
|
_reRankModelMap.set(model.name, model);
|
|
if (model.isDefault) {
|
|
_systemDefaultModel.rerank = model;
|
|
}
|
|
}
|
|
}
|
|
};
|
|
|
|
try {
|
|
// Get model from db and plugin
|
|
const [dbModels, systemModels] = await Promise.all([
|
|
MongoSystemModel.find({}).lean(),
|
|
pluginClient
|
|
.listModels()
|
|
.then((res) => res)
|
|
.catch(() => [])
|
|
]);
|
|
|
|
// Load system model from local
|
|
systemModels.forEach((model) => {
|
|
const mergeObject = (obj1: any, obj2: any) => {
|
|
if (!obj1 && !obj2) return undefined;
|
|
const formatObj1 = typeof obj1 === 'object' ? obj1 : {};
|
|
const formatObj2 = typeof obj2 === 'object' ? obj2 : {};
|
|
return { ...formatObj1, ...formatObj2 };
|
|
};
|
|
|
|
const dbModel = dbModels.find((item) => item.model === model.model);
|
|
const provider = getModelProvider(dbModel?.metadata?.provider || model.provider, language);
|
|
|
|
const modelData: any = {
|
|
...model,
|
|
...dbModel?.metadata,
|
|
provider: provider.id,
|
|
avatar: provider.avatar,
|
|
type: dbModel?.metadata?.type || model.type,
|
|
isCustom: false,
|
|
|
|
...(model.type === ModelTypeEnum.llm && {
|
|
maxResponse: model.maxTokens || 4000
|
|
}),
|
|
|
|
...(model.type === ModelTypeEnum.llm && dbModel?.metadata?.type === ModelTypeEnum.llm
|
|
? {
|
|
maxResponse: dbModel?.metadata?.maxResponse ?? model.maxTokens ?? 4000,
|
|
defaultConfig: mergeObject(model.defaultConfig, dbModel?.metadata?.defaultConfig),
|
|
fieldMap: mergeObject(model.fieldMap, dbModel?.metadata?.fieldMap),
|
|
maxTokens: undefined
|
|
}
|
|
: {})
|
|
};
|
|
pushModel(modelData);
|
|
});
|
|
|
|
// Custom model(Not in system config)
|
|
dbModels.forEach((dbModel) => {
|
|
if (_systemModelList.find((item) => item.model === dbModel.model)) return;
|
|
|
|
pushModel({
|
|
...dbModel.metadata,
|
|
isCustom: true
|
|
});
|
|
});
|
|
|
|
// Sort model list
|
|
_systemActiveModelList.sort((a, b) => {
|
|
const providerA = getModelProvider(a.provider, language);
|
|
const providerB = getModelProvider(b.provider, language);
|
|
return providerA.order - providerB.order;
|
|
});
|
|
|
|
// Default model check
|
|
{
|
|
if (!_systemDefaultModel.llm) {
|
|
_systemDefaultModel.llm = Array.from(_llmModelMap.values())[0];
|
|
}
|
|
if (!_systemDefaultModel.datasetTextLLM) {
|
|
_systemDefaultModel.datasetTextLLM = Array.from(_llmModelMap.values()).find(
|
|
(item) => item.datasetProcess
|
|
);
|
|
}
|
|
if (!_systemDefaultModel.datasetImageLLM) {
|
|
_systemDefaultModel.datasetImageLLM = Array.from(_llmModelMap.values()).find(
|
|
(item) => item.vision
|
|
);
|
|
}
|
|
if (!_systemDefaultModel.helperBotLLM) {
|
|
_systemDefaultModel.helperBotLLM = _systemActiveModelList.find(
|
|
(item) => item.type === ModelTypeEnum.llm
|
|
);
|
|
}
|
|
if (!_systemDefaultModel.embedding) {
|
|
_systemDefaultModel.embedding = Array.from(_embeddingModelMap.values())[0];
|
|
}
|
|
if (!_systemDefaultModel.tts) {
|
|
_systemDefaultModel.tts = Array.from(_ttsModelMap.values())[0];
|
|
}
|
|
if (!_systemDefaultModel.stt) {
|
|
_systemDefaultModel.stt = Array.from(_sttModelMap.values())[0];
|
|
}
|
|
if (!_systemDefaultModel.rerank) {
|
|
_systemDefaultModel.rerank = Array.from(_reRankModelMap.values())[0];
|
|
}
|
|
}
|
|
|
|
// Set global value
|
|
{
|
|
global.systemModelList = _systemModelList;
|
|
global.systemActiveModelList = _systemActiveModelList;
|
|
global.llmModelMap = _llmModelMap;
|
|
global.embeddingModelMap = _embeddingModelMap;
|
|
global.ttsModelMap = _ttsModelMap;
|
|
global.sttModelMap = _sttModelMap;
|
|
global.reRankModelMap = _reRankModelMap;
|
|
global.systemDefaultModel = _systemDefaultModel;
|
|
global.systemActiveDesensitizedModels = _systemActiveModelList.map((model) => ({
|
|
...model,
|
|
defaultSystemChatPrompt: undefined,
|
|
fieldMap: undefined,
|
|
defaultConfig: undefined,
|
|
weight: undefined,
|
|
dbConfig: undefined,
|
|
queryConfig: undefined,
|
|
requestUrl: undefined,
|
|
requestAuth: undefined
|
|
})) as SystemModelItemType[];
|
|
}
|
|
|
|
const logger = getLogger(LogCategories.MODULE.AI.CONFIG);
|
|
logger.debug('System models loaded', {
|
|
total: _systemModelList.length,
|
|
active: _systemActiveModelList.length
|
|
});
|
|
} catch (error) {
|
|
const logger = getLogger(LogCategories.MODULE.AI.CONFIG);
|
|
logger.error('System models load failed', { error });
|
|
|
|
return Promise.reject(error);
|
|
}
|
|
};
|
|
|
|
export const getSystemModelConfig = async (model: string): Promise<SystemModelItemType> => {
|
|
const modelData = findModelFromAlldata(model);
|
|
if (!modelData) return Promise.reject('Model is not found');
|
|
if (modelData.isCustom) return Promise.reject('Custom model not data');
|
|
|
|
// Read file
|
|
const modelDefaulConfig = await pluginClient
|
|
.listModels()
|
|
.then((models) => models.find((item) => item.model === model) as SystemModelItemType);
|
|
|
|
return {
|
|
...modelDefaulConfig,
|
|
provider: modelData.provider,
|
|
isCustom: false
|
|
};
|
|
};
|
|
|
|
export const watchSystemModelUpdate = () => {
|
|
const changeStream = MongoSystemModel.watch();
|
|
|
|
return changeStream.on(
|
|
'change',
|
|
debounce(async () => {
|
|
try {
|
|
// Main node will reload twice
|
|
await loadSystemModels(true);
|
|
// All node reaload buffer
|
|
await reloadFastGPTConfigBuffer();
|
|
} catch (error) {}
|
|
}, 500)
|
|
);
|
|
};
|
|
|
|
// 更新完模型后,需要重载缓存
|
|
export const updatedReloadSystemModel = async () => {
|
|
// 1. 更新模型(所有节点都会触发)
|
|
await loadSystemModels(true);
|
|
// 2. 更新缓存(仅主节点触发)
|
|
await updateFastGPTConfigBuffer();
|
|
await refreshVersionKey(SystemCacheKeyEnum.modelPermission, '*');
|
|
// 3. 延迟1秒,等待其他节点刷新
|
|
await delay(1000);
|
|
};
|
|
export const cronRefreshModels = async () => {
|
|
setCron('*/5 * * * *', async () => {
|
|
// 1. 更新模型(所有节点都会触发)
|
|
await loadSystemModels(true);
|
|
// 2. 更新缓存(仅主节点触发)
|
|
await updateFastGPTConfigBuffer();
|
|
});
|
|
};
|