mirror of
https://github.com/labring/FastGPT.git
synced 2025-07-23 05:12:39 +00:00

* Json completion (#16) * json-completion * fix duplicate * fix * fix: config json * feat: query extension * perf: i18n * 468 doc * json editor * perf: doc * perf: default extension model * docker file * doc * perf: token count * perf: search extension * format * perf: some constants data --------- Co-authored-by: heheer <71265218+newfish-cmyk@users.noreply.github.com>
86 lines
2.2 KiB
TypeScript
86 lines
2.2 KiB
TypeScript
/* Only the token of gpt-3.5-turbo is used */
|
|
import type { ChatItemType } from '../../../core/chat/type';
|
|
import { Tiktoken } from 'js-tiktoken/lite';
|
|
import { adaptChat2GptMessages } from '../../../core/chat/adapt';
|
|
import { ChatCompletionRequestMessageRoleEnum } from '../../../core/ai/constant';
|
|
import encodingJson from './cl100k_base.json';
|
|
|
|
/* init tikToken obj */
|
|
export function getTikTokenEnc() {
|
|
if (typeof window !== 'undefined' && window.TikToken) {
|
|
return window.TikToken;
|
|
}
|
|
if (typeof global !== 'undefined' && global.TikToken) {
|
|
return global.TikToken;
|
|
}
|
|
|
|
const enc = new Tiktoken(encodingJson);
|
|
|
|
if (typeof window !== 'undefined') {
|
|
window.TikToken = enc;
|
|
}
|
|
if (typeof global !== 'undefined') {
|
|
global.TikToken = enc;
|
|
}
|
|
|
|
return enc;
|
|
}
|
|
|
|
/* count one prompt tokens */
|
|
export function countPromptTokens(
|
|
prompt = '',
|
|
role: '' | `${ChatCompletionRequestMessageRoleEnum}` = ''
|
|
) {
|
|
const enc = getTikTokenEnc();
|
|
const text = `${role}\n${prompt}`;
|
|
|
|
try {
|
|
const encodeText = enc.encode(text);
|
|
return encodeText.length + role.length; // 补充 role 估算值
|
|
} catch (error) {
|
|
return text.length;
|
|
}
|
|
}
|
|
|
|
/* count messages tokens */
|
|
export function countMessagesTokens({ messages }: { messages: ChatItemType[] }) {
|
|
const adaptMessages = adaptChat2GptMessages({ messages, reserveId: true });
|
|
|
|
let totalTokens = 0;
|
|
for (let i = 0; i < adaptMessages.length; i++) {
|
|
const item = adaptMessages[i];
|
|
const tokens = countPromptTokens(item.content, item.role);
|
|
totalTokens += tokens;
|
|
}
|
|
|
|
return totalTokens;
|
|
}
|
|
|
|
/* slice messages from top to bottom by maxTokens */
|
|
export function sliceMessagesTB({
|
|
messages,
|
|
maxTokens
|
|
}: {
|
|
messages: ChatItemType[];
|
|
maxTokens: number;
|
|
}) {
|
|
const adaptMessages = adaptChat2GptMessages({ messages, reserveId: true });
|
|
let reduceTokens = maxTokens;
|
|
let result: ChatItemType[] = [];
|
|
|
|
for (let i = 0; i < adaptMessages.length; i++) {
|
|
const item = adaptMessages[i];
|
|
|
|
const tokens = countPromptTokens(item.content, item.role);
|
|
reduceTokens -= tokens;
|
|
|
|
if (reduceTokens > 0) {
|
|
result.push(messages[i]);
|
|
} else {
|
|
break;
|
|
}
|
|
}
|
|
|
|
return result.length === 0 && messages[0] ? [messages[0]] : result;
|
|
}
|