mirror of
https://github.com/labring/FastGPT.git
synced 2026-05-03 01:02:15 +08:00
4b24472106
* docs(i18n): translate batch 1 * docs(i18n): translate batch 2 * docs(i18n): translate batch 3 (20 files) - openapi/: app, share - faq/: all 8 files - use-cases/: index, external-integration (5 files), app-cases (4 files) Translated using North American style with natural, concise language. Preserved MDX syntax, code blocks, images, and component imports. * docs(i18n): translate protocol docs * docs(i18n): translate introduction docs (part 1) * docs(i18n): translate use-cases docs * docs(i18n): translate introduction docs (part 2 - batch 1) * docs(i18n): translate final 9 files * fix(i18n): fix YAML and MDX syntax errors in translated files - Add quotes to description with colon in submit_application_template.en.mdx - Remove duplicate Chinese content in translate-subtitle-using-gpt.en.mdx - Fix unclosed details tag issue * docs(i18n): translate all meta.json navigation files * fix(i18n): translate Chinese separators in meta.en.json files * translate * translate * i18n --------- Co-authored-by: archer <archer@archerdeMac-mini.local> Co-authored-by: archer <545436317@qq.com>
102 lines
3.4 KiB
Plaintext
102 lines
3.4 KiB
Plaintext
---
|
|
title: MCP Server
|
|
description: A quick overview of FastGPT MCP Server
|
|
---
|
|
|
|
## What is MCP Server?
|
|
|
|
MCP (Model Context Protocol) was released by Anthropic in early November 2024. It standardizes communication between AI models and external systems, simplifying integration. With OpenAI officially supporting MCP, more and more AI vendors are adopting the protocol.
|
|
|
|
MCP has two main components: Client and Server. The Client is the AI model consumer — it uses MCP Client to give the model the ability to call external systems. The Server provides and runs those external system integrations.
|
|
|
|
FastGPT's MCP Server feature lets you select `multiple` applications built on FastGPT and expose them via MCP protocol for external consumption.
|
|
|
|
Currently, FastGPT's MCP Server uses the SSE transport protocol, with plans to migrate to `HTTP Streamable` in the future.
|
|
|
|
## Using MCP Server in FastGPT
|
|
|
|
### 1. Create an MCP Server
|
|
|
|
After logging into FastGPT, open `Workspace` and click `MCP Server` to access the management page. Here you can see all your MCP Servers and the number of applications each one manages.
|
|
|
|

|
|
|
|
You can customize the MCP Server name and select which applications to associate.
|
|
|
|
| | |
|
|
|---|---|
|
|
|  |  |
|
|
|
|
### 2. Get the MCP Server URL
|
|
|
|
After creating an MCP Server, click `Start Using` to get the access URL.
|
|
|
|
| | |
|
|
|---|---|
|
|
|  |  |
|
|
|
|
#### 3. Use the MCP Server
|
|
|
|
Use the URL in any MCP-compatible client to call your FastGPT applications — for example, `Cursor` or `Cherry Studio`. Here's how to set it up in Cursor.
|
|
|
|
Open Cursor's settings page and click MCP to enter the MCP configuration page. Click the new MCP Server button to open a JSON configuration file. Paste the `integration script` from step 2 into the `JSON file` and save.
|
|
|
|
Return to Cursor's MCP management page and you'll see your MCP Server listed. Make sure to set it to `enabled`.
|
|
|
|
| | | |
|
|
|---|---|---|
|
|
|  |  |  |
|
|
|
|
|
|
Open Cursor's chat panel and switch to `Agent` mode — only this mode triggers MCP Server calls.
|
|
After sending a question about `fastgpt`, you'll see Cursor invoke an MCP tool (described as: query fastgpt knowledge base), which calls the FastGPT application to process the question and return results.
|
|
|
|
| | |
|
|
|---|---|
|
|
|  |  |
|
|
|
|
|
|
## Self-Hosted MCP Server Setup
|
|
|
|
Self-hosted FastGPT deployments require version `v4.9.6` or higher to use MCP Server.
|
|
|
|
### Update docker-compose.yml
|
|
|
|
Add the `fastgpt-mcp-server` service to your `docker-compose.yml`:
|
|
|
|
```yml
|
|
fastgpt-mcp-server:
|
|
container_name: fastgpt-mcp-server
|
|
image: ghcr.io/labring/fastgpt-mcp_server:latest
|
|
ports:
|
|
- 3005:3000
|
|
networks:
|
|
- fastgpt
|
|
restart: always
|
|
environment:
|
|
- FASTGPT_ENDPOINT=http://fastgpt:3000
|
|
```
|
|
|
|
### Update FastGPT Configuration
|
|
|
|
In your `config.json`, add: `"feconfigs.mcpServerProxyEndpoint": "<your fastgpt-mcp-server URL>"` (no trailing slash). For example:
|
|
```json
|
|
{
|
|
"feConfigs": {
|
|
"lafEnv": "https://laf.dev",
|
|
"mcpServerProxyEndpoint": "https://mcp.fastgpt.cn"
|
|
}
|
|
}
|
|
```
|
|
|
|
### Restart FastGPT
|
|
|
|
Since you modified a mounted config file, force a full restart:
|
|
|
|
```bash
|
|
docker-compose down
|
|
docker-compose up -d
|
|
```
|
|
|
|
After restarting, the MCP Server option will appear in the Workspace.
|