mirror of
https://github.com/labring/FastGPT.git
synced 2026-05-10 01:08:08 +08:00
76 lines
3.0 KiB
Plaintext
76 lines
3.0 KiB
Plaintext
---
|
|
title: Knowledge Base Usage
|
|
description: Common Knowledge Base usage questions
|
|
---
|
|
|
|
## Uploaded file content shows garbled characters
|
|
|
|
Re-save the file with UTF-8 encoding.
|
|
|
|
## What's the difference between the File Processing Model and the Index Model in Knowledge Base settings?
|
|
|
|
* **File Processing Model**: Used for **Enhanced Processing** and **Q&A Splitting** during data ingestion. Enhanced Processing generates related questions and summaries; Q&A Splitting generates question-answer pairs.
|
|
* **Index Model**: Used for vectorization — it processes and organizes text data into a structure optimized for fast retrieval.
|
|
|
|
## Does the Knowledge Base support Excel files?
|
|
|
|
Yes. You can upload xlsx and other spreadsheet formats, not just CSV.
|
|
|
|
## How are Knowledge Base tokens calculated?
|
|
|
|
All token counts use the GPT-3.5 tokenizer as the standard.
|
|
|
|
## I accidentally deleted the rerank model. How do I add it back?
|
|
|
|

|
|
|
|
Add the rerank model configuration in your `config.json` file, then you'll be able to select it again.
|
|
|
|
## If I created apps and Knowledge Bases on the cloud platform, will my data be deleted if I don't renew right away?
|
|
|
|
On the free plan, Knowledge Base data is cleared after 30 days of inactivity (no login). Apps are not affected. Paid plans automatically downgrade to the free plan upon expiration.
|
|

|
|
|
|
## The AI stops responding mid-answer when there are too many relevant Knowledge Base results.
|
|
|
|
FastGPT calculates the maximum response length as:
|
|
|
|
Max Response = min(Configured Max Response, Max Context Window - History)
|
|
|
|
For example, with an 18K context model, input + output share the same window. As output grows, available input shrinks.
|
|
|
|
To fix this:
|
|
|
|
1. Check your configured max response (response limit) setting.
|
|
2. Reduce input to free up space for output — specifically, reduce the number of chat history turns included in the workflow.
|
|
|
|
Where to find the max response setting:
|
|
|
|

|
|
|
|

|
|
|
|
For self-hosted deployments, you can reserve headroom when configuring model context limits. For example, set a 128K model to 120K — the remaining space will be allocated to output.
|
|
|
|
|
|
## I'm hitting context limit errors before reaching the configured number of chat history turns.
|
|
|
|
FastGPT calculates the maximum response length as:
|
|
|
|
Max Response = min(Configured Max Response, Max Context Window - History)
|
|
|
|
For example, with an 18K context model, input + output share the same window. As output grows, available input shrinks.
|
|
|
|
To fix this:
|
|
|
|
1. Check your configured max response (response limit) setting.
|
|
2. Reduce input to free up space for output — specifically, reduce the number of chat history turns included in the workflow.
|
|
|
|
Where to find the max response setting:
|
|
|
|

|
|
|
|

|
|
|
|
For self-hosted deployments, you can reserve headroom when configuring model context limits. For example, set a 128K model to 120K — the remaining space will be allocated to output.
|