mirror of
https://github.com/labring/FastGPT.git
synced 2025-07-22 20:37:48 +00:00

* update: Add type * fix: update import statement for NextApiRequest type * fix: update imports to use type for LexicalEditor and EditorState * Refactor imports to use 'import type' for type-only imports across multiple files - Updated imports in various components and API files to use 'import type' for better clarity and to optimize TypeScript's type checking. - Ensured consistent usage of type imports in files related to chat, dataset, workflow, and user management. - Improved code readability and maintainability by distinguishing between value and type imports. * refactor: remove old ESLint configuration and add new rules - Deleted the old ESLint configuration file from the app project. - Added a new ESLint configuration file with updated rules and settings. - Changed imports to use type-only imports in various files for better clarity and performance. - Updated TypeScript configuration to remove unnecessary options. - Added an ESLint ignore file to exclude build and dependency directories from linting. * fix: update imports to use 'import type' for type-only imports in schema files
225 lines
6.7 KiB
TypeScript
225 lines
6.7 KiB
TypeScript
import { batchRun, delay } from '@fastgpt/global/common/system/utils';
|
|
import { addLog } from '../../common/system/log';
|
|
import { htmlTable2Md } from '@fastgpt/global/common/string/markdown';
|
|
import axios, { type Method } from 'axios';
|
|
import { getNanoid } from '@fastgpt/global/common/string/tools';
|
|
import { getErrText } from '@fastgpt/global/common/error/utils';
|
|
import { type ImageType } from '../../worker/readFile/type';
|
|
import { getImageBase64 } from '../../common/file/image/utils';
|
|
|
|
type ApiResponseDataType<T = any> = {
|
|
code: string;
|
|
msg?: string;
|
|
data: T;
|
|
};
|
|
|
|
export const useDoc2xServer = ({ apiKey }: { apiKey: string }) => {
|
|
// Init request
|
|
const instance = axios.create({
|
|
baseURL: 'https://v2.doc2x.noedgeai.com/api',
|
|
timeout: 60000,
|
|
headers: {
|
|
Authorization: `Bearer ${apiKey}`
|
|
}
|
|
});
|
|
// Response check
|
|
const checkRes = (data: ApiResponseDataType) => {
|
|
if (data === undefined) {
|
|
addLog.info('[Doc2x] Server data is empty');
|
|
return Promise.reject('服务器异常');
|
|
}
|
|
return data;
|
|
};
|
|
const responseError = (err: any) => {
|
|
if (!err) {
|
|
return Promise.reject({ message: '[Doc2x] Unknown error' });
|
|
}
|
|
if (typeof err === 'string') {
|
|
return Promise.reject({ message: `[Doc2x] ${err}` });
|
|
}
|
|
if (typeof err.message === 'string') {
|
|
return Promise.reject({ message: `[Doc2x] ${err.message}` });
|
|
}
|
|
if (typeof err.data === 'string') {
|
|
return Promise.reject({ message: `[Doc2x] ${err.data}` });
|
|
}
|
|
if (err?.response?.data) {
|
|
return Promise.reject({ message: `[Doc2x] ${getErrText(err?.response?.data)}` });
|
|
}
|
|
|
|
addLog.error('[Doc2x] Unknown error', err);
|
|
return Promise.reject({ message: `[Doc2x] ${getErrText(err)}` });
|
|
};
|
|
const request = <T>(url: string, data: any, method: Method): Promise<ApiResponseDataType<T>> => {
|
|
// Remove empty data
|
|
for (const key in data) {
|
|
if (data[key] === undefined) {
|
|
delete data[key];
|
|
}
|
|
}
|
|
|
|
return instance
|
|
.request({
|
|
url,
|
|
method,
|
|
data: ['POST', 'PUT'].includes(method) ? data : undefined,
|
|
params: !['POST', 'PUT'].includes(method) ? data : undefined
|
|
})
|
|
.then((res) => checkRes(res.data))
|
|
.catch((err) => responseError(err));
|
|
};
|
|
|
|
const parsePDF = async (fileBuffer: Buffer) => {
|
|
addLog.debug('[Doc2x] PDF parse start');
|
|
const startTime = Date.now();
|
|
|
|
// 1. Get pre-upload URL first
|
|
const {
|
|
code,
|
|
msg,
|
|
data: preupload_data
|
|
} = await request<{ uid: string; url: string }>('/v2/parse/preupload', null, 'POST');
|
|
if (!['ok', 'success'].includes(code)) {
|
|
return Promise.reject(`[Doc2x] Failed to get pre-upload URL: ${msg}`);
|
|
}
|
|
const upload_url = preupload_data.url;
|
|
const uid = preupload_data.uid;
|
|
|
|
// 2. Upload file to pre-signed URL with binary stream
|
|
const blob = new Blob([fileBuffer], { type: 'application/pdf' });
|
|
const response = await axios
|
|
.put(upload_url, blob, {
|
|
headers: {
|
|
'Content-Type': 'application/pdf'
|
|
}
|
|
})
|
|
.catch((error) => {
|
|
return Promise.reject(`[Doc2x] Failed to upload file: ${getErrText(error)}`);
|
|
});
|
|
if (response.status !== 200) {
|
|
return Promise.reject(
|
|
`[Doc2x] Upload failed with status ${response.status}: ${response.statusText}`
|
|
);
|
|
}
|
|
addLog.debug(`[Doc2x] Uploaded file success, uid: ${uid}`);
|
|
|
|
await delay(5000);
|
|
|
|
// 3. Get the result by uid
|
|
const checkResult = async () => {
|
|
// 10 minutes
|
|
let retry = 120;
|
|
|
|
while (retry > 0) {
|
|
try {
|
|
const {
|
|
code,
|
|
data: result_data,
|
|
msg
|
|
} = await request<{
|
|
progress: number;
|
|
status: 'processing' | 'failed' | 'success';
|
|
result: {
|
|
pages: {
|
|
md: string;
|
|
}[];
|
|
};
|
|
}>(`/v2/parse/status?uid=${uid}`, null, 'GET');
|
|
|
|
// Error
|
|
if (!['ok', 'success'].includes(code)) {
|
|
return Promise.reject(`[Doc2x] Failed to get result (uid: ${uid}): ${msg}`);
|
|
}
|
|
|
|
// Process
|
|
if (['ready', 'processing'].includes(result_data.status)) {
|
|
addLog.debug(`[Doc2x] Waiting for the result, uid: ${uid}`);
|
|
await delay(5000);
|
|
}
|
|
|
|
// Finifsh
|
|
if (result_data.status === 'success') {
|
|
return {
|
|
text: result_data.result.pages
|
|
.map((page) => page.md)
|
|
.join('')
|
|
.replace(/\\[\(\)]/g, '$')
|
|
.replace(/\\[\[\]]/g, '$$')
|
|
.replace(/<img\s+src="([^"]+)"(?:\s*\?[^>]*)?(?:\s*\/>|>)/g, '')
|
|
.replace(/<!-- Media -->/g, '')
|
|
.replace(/<!-- Footnote -->/g, '')
|
|
.replace(/\$(.+?)\s+\\tag\{(.+?)\}\$/g, '$$$1 \\qquad \\qquad ($2)$$')
|
|
.replace(/\\text\{([^}]*?)(\b\w+)_(\w+\b)([^}]*?)\}/g, '\\text{$1$2\\_$3$4}'),
|
|
pages: result_data.result.pages.length
|
|
};
|
|
}
|
|
} catch (error) {
|
|
// Just network error
|
|
addLog.warn(`[Doc2x] Get result error`, { error });
|
|
await delay(500);
|
|
}
|
|
|
|
retry--;
|
|
}
|
|
return Promise.reject(`[Doc2x] Failed to get result (uid: ${uid}): Process timeout`);
|
|
};
|
|
|
|
const { text, pages } = await checkResult();
|
|
|
|
//  => 
|
|
const parseTextImage = async (text: string) => {
|
|
// Extract image links and convert to base64
|
|
const imageList: { id: string; url: string }[] = [];
|
|
let processedText = text.replace(/!\[.*?\]\((http[^)]+)\)/g, (match, url) => {
|
|
const id = `IMAGE_${getNanoid()}_IMAGE`;
|
|
imageList.push({
|
|
id,
|
|
url
|
|
});
|
|
return ``;
|
|
});
|
|
|
|
// Get base64 from image url
|
|
let resultImageList: ImageType[] = [];
|
|
await batchRun(
|
|
imageList,
|
|
async (item) => {
|
|
try {
|
|
const { base64, mime } = await getImageBase64(item.url);
|
|
resultImageList.push({
|
|
uuid: item.id,
|
|
mime,
|
|
base64
|
|
});
|
|
} catch (error) {
|
|
processedText = processedText.replace(item.id, item.url);
|
|
addLog.warn(`[Doc2x] Failed to get image from ${item.url}: ${getErrText(error)}`);
|
|
}
|
|
},
|
|
5
|
|
);
|
|
|
|
return {
|
|
text: processedText,
|
|
imageList: resultImageList
|
|
};
|
|
};
|
|
const { text: formatText, imageList } = await parseTextImage(htmlTable2Md(text));
|
|
|
|
addLog.debug(`[Doc2x] PDF parse finished`, {
|
|
time: `${Math.round((Date.now() - startTime) / 1000)}s`,
|
|
pages
|
|
});
|
|
|
|
return {
|
|
pages,
|
|
text: formatText,
|
|
imageList
|
|
};
|
|
};
|
|
|
|
return {
|
|
parsePDF
|
|
};
|
|
};
|