mirror of
https://github.com/labring/FastGPT.git
synced 2025-07-23 21:13:50 +00:00

* perf: plan tip * perf: upload size controller * feat: add image ttl index * feat: new upload file ux * remove file * feat: support read pptx * feat: support xlsx * fix: rerank docker flie
58 lines
1.3 KiB
TypeScript
58 lines
1.3 KiB
TypeScript
import { DatasetTrainingSchemaType } from '@fastgpt/global/core/dataset/type';
|
|
import { addLog } from '../../../common/system/log';
|
|
import { getErrText } from '@fastgpt/global/common/error/utils';
|
|
import { MongoDatasetTraining } from './schema';
|
|
import Papa from 'papaparse';
|
|
|
|
export const checkInvalidChunkAndLock = async ({
|
|
err,
|
|
errText,
|
|
data
|
|
}: {
|
|
err: any;
|
|
errText: string;
|
|
data: DatasetTrainingSchemaType;
|
|
}) => {
|
|
if (err?.response) {
|
|
addLog.error(`openai error: ${errText}`, {
|
|
status: err.response?.status,
|
|
statusText: err.response?.statusText,
|
|
data: err.response?.data
|
|
});
|
|
} else {
|
|
addLog.error(getErrText(err, errText), err);
|
|
}
|
|
|
|
if (
|
|
err?.message === 'invalid message format' ||
|
|
err?.type === 'invalid_request_error' ||
|
|
err?.code === 500
|
|
) {
|
|
addLog.info('Lock training data');
|
|
console.log(err);
|
|
|
|
try {
|
|
await MongoDatasetTraining.findByIdAndUpdate(data._id, {
|
|
lockTime: new Date('2998/5/5')
|
|
});
|
|
} catch (error) {}
|
|
return true;
|
|
}
|
|
return false;
|
|
};
|
|
|
|
export const parseCsvTable2Chunks = (rawText: string) => {
|
|
const csvArr = Papa.parse(rawText).data as string[][];
|
|
|
|
const chunks = csvArr
|
|
.map((item) => ({
|
|
q: item[0] || '',
|
|
a: item[1] || ''
|
|
}))
|
|
.filter((item) => item.q || item.a);
|
|
|
|
return {
|
|
chunks
|
|
};
|
|
};
|