mirror of
https://github.com/labring/FastGPT.git
synced 2026-02-27 01:02:22 +08:00
* refactor: fastgpt object storage & global proxy (#6155) * feat: migrate to fastgpt storage sdk * chore: rename env variable * chore: move to sdk dir * docs: object storage * CHORE * chore: storage mocks * chore: update docker-compose * fix: global proxy agent * fix: update COS proxy * refactor: use fetch instead of http.request * fix: axios request base url * fix: axios proxy request behavior * fix: bumps axios * fix: patch axios for proxy * fix: replace axios with proxied axios * fix: upload txt file encoding * clean code * fix: use "minio" for minio adapter (#6205) * fix: use minio client to delete files when using minio vendor (#6206) * doc * feat: filter citations and add response button control (#6170) * feat: filter citations and add response button control * i18n * fix * fix test * perf: chat api code * fix: workflow edge overlap and auto-align in folded loop nodes (#6204) * fix: workflow edge overlap and auto-align in folded loop nodes * sort * fix * fix edge * fix icon * perf: s3 file name * perf: admin get app api * perf: catch user error * fix: refactor useOrg hook to use debounced search key (#6180) * chore: comment minio adapter (#6207) * chore: filename with suffix random id * perf: s3 storage code * fix: encode filename when copy object --------- Co-authored-by: archer <545436317@qq.com> * fix: node card link * json * perf: chat index; * index * chat item soft delete (#6216) * chat item soft delete * temp * fix * remove code * perf: delete chat item --------- Co-authored-by: archer <545436317@qq.com> * feat: select wheather filter sensitive info when export apps (#6222) * fix some bugs (#6210) * fix v4.14.5 bugs * type * fix * fix * custom feedback * fix * code * fix * remove invalid function --------- Co-authored-by: archer <545436317@qq.com> * perf: test * fix file default local upload (#6223) * docs: improve object storage introduction (#6224) * doc --------- Co-authored-by: roy <whoeverimf5@gmail.com> Co-authored-by: heheer <heheer@sealos.io> Co-authored-by: Finley Ge <32237950+FinleyGe@users.noreply.github.com>
237 lines
7.2 KiB
TypeScript
237 lines
7.2 KiB
TypeScript
import { batchRun, delay } from '@fastgpt/global/common/system/utils';
|
|
import { addLog } from '../../common/system/log';
|
|
import { htmlTable2Md } from '@fastgpt/global/common/string/markdown';
|
|
import { type Method } from 'axios';
|
|
import { getNanoid } from '@fastgpt/global/common/string/tools';
|
|
import { getErrText } from '@fastgpt/global/common/error/utils';
|
|
import { type ImageType } from '../../worker/readFile/type';
|
|
import { getImageBase64 } from '../../common/file/image/utils';
|
|
import { createProxyAxios, axios } from '../../common/api/axios';
|
|
|
|
type ApiResponseDataType<T = any> = {
|
|
code: string;
|
|
msg?: string;
|
|
data: T;
|
|
};
|
|
|
|
export const useDoc2xServer = ({ apiKey }: { apiKey: string }) => {
|
|
// Init request
|
|
const instance = createProxyAxios({
|
|
baseURL: 'https://v2.doc2x.noedgeai.com/api',
|
|
timeout: 60000,
|
|
headers: {
|
|
Authorization: `Bearer ${apiKey}`
|
|
}
|
|
});
|
|
// Response check
|
|
const checkRes = (data: ApiResponseDataType) => {
|
|
if (data === undefined) {
|
|
addLog.info('[Doc2x] Server data is empty');
|
|
return Promise.reject('服务器异常');
|
|
}
|
|
return data;
|
|
};
|
|
const responseError = (err: any) => {
|
|
if (!err) {
|
|
return Promise.reject({ message: '[Doc2x] Unknown error' });
|
|
}
|
|
if (typeof err === 'string') {
|
|
return Promise.reject({ message: `[Doc2x] ${err}` });
|
|
}
|
|
if (typeof err.data === 'string') {
|
|
return Promise.reject({ message: `[Doc2x] ${err.data}` });
|
|
}
|
|
if (err?.response?.data) {
|
|
return Promise.reject({ message: `[Doc2x] ${getErrText(err?.response?.data)}` });
|
|
}
|
|
if (typeof err.message === 'string') {
|
|
return Promise.reject({ message: `[Doc2x] ${err.message}` });
|
|
}
|
|
|
|
addLog.error('[Doc2x] Unknown error', err);
|
|
return Promise.reject({ message: `[Doc2x] ${getErrText(err)}` });
|
|
};
|
|
const request = <T>(url: string, data: any, method: Method): Promise<ApiResponseDataType<T>> => {
|
|
// Remove empty data
|
|
for (const key in data) {
|
|
if (data[key] === undefined) {
|
|
delete data[key];
|
|
}
|
|
}
|
|
|
|
return instance
|
|
.request({
|
|
url,
|
|
method,
|
|
data: ['POST', 'PUT'].includes(method) ? data : undefined,
|
|
params: !['POST', 'PUT'].includes(method) ? data : undefined
|
|
})
|
|
.then((res) => checkRes(res.data))
|
|
.catch((err) => responseError(err));
|
|
};
|
|
|
|
const parsePDF = async (fileBuffer: Buffer) => {
|
|
addLog.debug('[Doc2x] PDF parse start');
|
|
const startTime = Date.now();
|
|
|
|
// 1. Get pre-upload URL first
|
|
const {
|
|
code,
|
|
msg,
|
|
data: preupload_data
|
|
} = await request<{ uid: string; url: string }>('/v2/parse/preupload', {}, 'POST');
|
|
if (!['ok', 'success'].includes(code)) {
|
|
return Promise.reject(`[Doc2x] Failed to get pre-upload URL: ${msg}`);
|
|
}
|
|
const upload_url = preupload_data.url;
|
|
const uid = preupload_data.uid;
|
|
|
|
// 2. Upload file to pre-signed URL with binary stream
|
|
const response = await axios
|
|
.put(upload_url, fileBuffer, {
|
|
headers: {
|
|
'Content-Type': 'application/pdf',
|
|
'Content-Length': fileBuffer.length.toString()
|
|
}
|
|
})
|
|
.catch((error) => {
|
|
return Promise.reject(`[Doc2x] Failed to upload file: ${getErrText(error)}`);
|
|
});
|
|
|
|
if (response.status !== 200) {
|
|
return Promise.reject(
|
|
`[Doc2x] Upload failed with status ${response.status}: ${response.statusText}`
|
|
);
|
|
}
|
|
addLog.debug(`[Doc2x] Uploaded file success, uid: ${uid}`);
|
|
|
|
await delay(5000);
|
|
|
|
// 3. Get the result by uid
|
|
const checkResult = async () => {
|
|
// 10 minutes
|
|
let retry = 120;
|
|
|
|
while (retry > 0) {
|
|
try {
|
|
const {
|
|
code,
|
|
data: result_data,
|
|
msg
|
|
} = await request<{
|
|
progress: number;
|
|
status: 'processing' | 'failed' | 'success';
|
|
result: {
|
|
pages: {
|
|
md: string;
|
|
}[];
|
|
};
|
|
}>(`/v2/parse/status?uid=${uid}`, null, 'GET');
|
|
|
|
// Error
|
|
if (!['ok', 'success'].includes(code)) {
|
|
return Promise.reject(`[Doc2x] Failed to get result (uid: ${uid}): ${msg}`);
|
|
}
|
|
|
|
// Process
|
|
if (['ready', 'processing'].includes(result_data.status)) {
|
|
addLog.debug(`[Doc2x] Waiting for the result, uid: ${uid}`);
|
|
await delay(5000);
|
|
}
|
|
|
|
// Finifsh
|
|
if (result_data.status === 'success') {
|
|
const cleanedText = result_data.result.pages
|
|
.map((page) => page.md)
|
|
.join('')
|
|
.replace(/\\[\(\)]/g, '$')
|
|
.replace(/\\[\[\]]/g, '$$')
|
|
.replace(/<img\s+src="([^"]+)"(?:\s*\?[^>]*)?(?:\s*\/>|>)/g, '')
|
|
.replace(/<!-- Media -->/g, '')
|
|
.replace(/<!-- Footnote -->/g, '')
|
|
.replace(/<!-- Meanless:[\s\S]*?-->/g, '')
|
|
.replace(/<!-- figureText:[\s\S]*?-->/g, '')
|
|
.replace(/\$(.+?)\s+\\tag\{(.+?)\}\$/g, '$$$1 \\qquad \\qquad ($2)$$')
|
|
.replace(/\\text\{([^}]*?)(\b\w+)_(\w+\b)([^}]*?)\}/g, '\\text{$1$2\\_$3$4}');
|
|
const remainingTags = cleanedText.match(/<!--[\s\S]*?-->/g);
|
|
if (remainingTags) {
|
|
addLog.warn(`[Doc2x] Remaining dirty tags after cleaning:`, {
|
|
count: remainingTags.length,
|
|
tags: remainingTags.slice(0, 3)
|
|
});
|
|
}
|
|
return {
|
|
text: cleanedText,
|
|
pages: result_data.result.pages.length
|
|
};
|
|
}
|
|
} catch (error) {
|
|
// Just network error
|
|
addLog.warn(`[Doc2x] Get result error`, { error });
|
|
await delay(500);
|
|
}
|
|
|
|
retry--;
|
|
}
|
|
return Promise.reject(`[Doc2x] Failed to get result (uid: ${uid}): Process timeout`);
|
|
};
|
|
|
|
const { text, pages } = await checkResult();
|
|
|
|
//  => 
|
|
const parseTextImage = async (text: string) => {
|
|
// Extract image links and convert to base64
|
|
const imageList: { id: string; url: string }[] = [];
|
|
let processedText = text.replace(/!\[.*?\]\((http[^)]+)\)/g, (match, url) => {
|
|
const id = `IMAGE_${getNanoid()}_IMAGE`;
|
|
imageList.push({
|
|
id,
|
|
url
|
|
});
|
|
return ``;
|
|
});
|
|
|
|
// Get base64 from image url
|
|
let resultImageList: ImageType[] = [];
|
|
await batchRun(
|
|
imageList,
|
|
async (item) => {
|
|
try {
|
|
const { base64, mime } = await getImageBase64(item.url);
|
|
resultImageList.push({
|
|
uuid: item.id,
|
|
mime,
|
|
base64
|
|
});
|
|
} catch (error) {
|
|
processedText = processedText.replace(item.id, item.url);
|
|
addLog.warn(`[Doc2x] Failed to get image from ${item.url}: ${getErrText(error)}`);
|
|
}
|
|
},
|
|
5
|
|
);
|
|
|
|
return {
|
|
text: processedText,
|
|
imageList: resultImageList
|
|
};
|
|
};
|
|
const { text: formatText, imageList } = await parseTextImage(htmlTable2Md(text));
|
|
|
|
addLog.debug(`[Doc2x] PDF parse finished`, {
|
|
time: `${Math.round((Date.now() - startTime) / 1000)}s`,
|
|
pages
|
|
});
|
|
|
|
return {
|
|
pages,
|
|
text: formatText,
|
|
imageList
|
|
};
|
|
};
|
|
|
|
return {
|
|
parsePDF
|
|
};
|
|
};
|