mirror of
https://github.com/labring/FastGPT.git
synced 2025-10-20 10:45:52 +00:00

* Revert "lafAccount add pat & re request when token invalid (#76)" (#77) This reverts commit 83d85dfe37adcaef4833385ea52ee79fd84720be. * perf: workflow ux * system config * Newflow (#89) * docs: Add doc for Xinference (#1266) Signed-off-by: Carson Yang <yangchuansheng33@gmail.com> * Revert "lafAccount add pat & re request when token invalid (#76)" (#77) This reverts commit 83d85dfe37adcaef4833385ea52ee79fd84720be. * perf: workflow ux * system config * Revert "lafAccount add pat & re request when token invalid (#76)" (#77) This reverts commit 83d85dfe37adcaef4833385ea52ee79fd84720be. * Revert "lafAccount add pat & re request when token invalid (#76)" (#77) This reverts commit 83d85dfe37adcaef4833385ea52ee79fd84720be. * Revert "lafAccount add pat & re request when token invalid (#76)" (#77) This reverts commit 83d85dfe37adcaef4833385ea52ee79fd84720be. * rename code * move code * update flow * input type selector * perf: workflow runtime * feat: node adapt newflow * feat: adapt plugin * feat: 360 connection * check workflow * perf: flow 性能 * change plugin input type (#81) * change plugin input type * plugin label mode * perf: nodecard * debug * perf: debug ui * connection ui * change workflow ui (#82) * feat: workflow debug * adapt openAPI for new workflow (#83) * adapt openAPI for new workflow * i18n * perf: plugin debug * plugin input ui * delete * perf: global variable select * fix rebase * perf: workflow performance * feat: input render type icon * input icon * adapt flow (#84) * adapt newflow * temp * temp * fix * feat: app schedule trigger * feat: app schedule trigger * perf: schedule ui * feat: ioslatevm run js code * perf: workflow varialbe table ui * feat: adapt simple mode * feat: adapt input params * output * feat: adapt tamplate * fix: ts * add if-else module (#86) * perf: worker * if else node * perf: tiktoken worker * fix: ts * perf: tiktoken * fix if-else node (#87) * fix if-else node * type * fix * perf: audio render * perf: Parallel worker * log * perf: if else node * adapt plugin * prompt * perf: reference ui * reference ui * handle ux * template ui and plugin tool * adapt v1 workflow * adapt v1 workflow completions * perf: time variables * feat: workflow keyboard shortcuts * adapt v1 workflow * update workflow example doc (#88) * fix: simple mode select tool --------- Signed-off-by: Carson Yang <yangchuansheng33@gmail.com> Co-authored-by: Carson Yang <yangchuansheng33@gmail.com> Co-authored-by: heheer <71265218+newfish-cmyk@users.noreply.github.com> * doc * perf: extract node * extra node field * update plugin version * doc * variable * change doc & fix prompt editor (#90) * fold workflow code * value type label --------- Signed-off-by: Carson Yang <yangchuansheng33@gmail.com> Co-authored-by: Carson Yang <yangchuansheng33@gmail.com> Co-authored-by: heheer <71265218+newfish-cmyk@users.noreply.github.com>
93 lines
2.4 KiB
TypeScript
93 lines
2.4 KiB
TypeScript
import {
|
|
ChatCompletionContentPart,
|
|
ChatCompletionCreateParams,
|
|
ChatCompletionMessageParam,
|
|
ChatCompletionTool
|
|
} from '@fastgpt/global/core/ai/type';
|
|
import { chats2GPTMessages } from '@fastgpt/global/core/chat/adapt';
|
|
import { ChatItemType } from '@fastgpt/global/core/chat/type';
|
|
import { WorkerNameEnum, getWorker } from '../../../worker/utils';
|
|
import { ChatCompletionRequestMessageRoleEnum } from '@fastgpt/global/core/ai/constants';
|
|
import { getNanoid } from '@fastgpt/global/common/string/tools';
|
|
import { addLog } from '../../system/log';
|
|
|
|
export const getTiktokenWorker = () => {
|
|
if (global.tiktokenWorker) {
|
|
return global.tiktokenWorker;
|
|
}
|
|
|
|
const worker = getWorker(WorkerNameEnum.countGptMessagesTokens);
|
|
|
|
worker.on('message', ({ id, data }: { id: string; data: number }) => {
|
|
const callback = global.tiktokenWorker?.callbackMap?.[id];
|
|
|
|
if (callback) {
|
|
callback?.(data);
|
|
delete global.tiktokenWorker.callbackMap[id];
|
|
}
|
|
});
|
|
|
|
global.tiktokenWorker = {
|
|
worker,
|
|
callbackMap: {}
|
|
};
|
|
|
|
return global.tiktokenWorker;
|
|
};
|
|
|
|
export const countGptMessagesTokens = (
|
|
messages: ChatCompletionMessageParam[],
|
|
tools?: ChatCompletionTool[],
|
|
functionCall?: ChatCompletionCreateParams.Function[]
|
|
) => {
|
|
return new Promise<number>((resolve) => {
|
|
const start = Date.now();
|
|
|
|
const { worker, callbackMap } = getTiktokenWorker();
|
|
const id = getNanoid();
|
|
|
|
const timer = setTimeout(() => {
|
|
resolve(0);
|
|
delete callbackMap[id];
|
|
}, 300);
|
|
|
|
callbackMap[id] = (data) => {
|
|
resolve(data);
|
|
clearTimeout(timer);
|
|
|
|
// 检测是否有内存泄漏
|
|
addLog.info(`Count token time: ${Date.now() - start}, token: ${data}`);
|
|
console.log(Object.keys(global.tiktokenWorker.callbackMap));
|
|
};
|
|
|
|
worker.postMessage({
|
|
id,
|
|
messages,
|
|
tools,
|
|
functionCall
|
|
});
|
|
});
|
|
};
|
|
|
|
export const countMessagesTokens = (messages: ChatItemType[]) => {
|
|
const adaptMessages = chats2GPTMessages({ messages, reserveId: true });
|
|
|
|
return countGptMessagesTokens(adaptMessages);
|
|
};
|
|
|
|
/* count one prompt tokens */
|
|
export const countPromptTokens = async (
|
|
prompt: string | ChatCompletionContentPart[] | null | undefined = '',
|
|
role: '' | `${ChatCompletionRequestMessageRoleEnum}` = ''
|
|
) => {
|
|
const total = await countGptMessagesTokens([
|
|
{
|
|
//@ts-ignore
|
|
role,
|
|
content: prompt
|
|
}
|
|
]);
|
|
|
|
return total;
|
|
};
|