mirror of
https://github.com/labring/FastGPT.git
synced 2025-11-29 01:05:13 +08:00
* feat: concat usage code (#5657) * feat: dataset parse queue (#5661) * feat: chat usage concat (#5669) * perf: search test usage * feat: chat usage concat * fix: ts * fix: ts * feat: chat node response store (#5675) * feat: chat node response store * limit export * test * add ai generate node (#5506) * add node copilot * apply code * update dynamic input & output * add code test * usage * dynamic input border render * optimize input & output * optimize code * update style * change card to popover * prompt editor basic * prompt editor * handle key down * update prompt * merge * fix * fix * fix * perf: workflow performance (#5677) * feat: chat node response store * limit export * perf: workflow performance * remove log * fix: app template get duplicate (#5682) * fix: dynamic input lock & code param (#5680) * fix: dynamic input lock & code param * fix * fix * feat: multi node data sync & system tool hot-swapping (#5575) * Enhance file upload functionality and system tool integration (#5257) * Enhance file upload functionality and system tool integration * Add supplementary documents and optimize the upload interface * Refactor file plugin types and update upload configurations * Refactor MinIO configuration variables and clean up API plugin handlers for improved readability and consistency * File name change * Refactor SystemTools component layout * fix i18n * fix * fix * fix * optimize app logs sort (#5310) * log keys config modal * multiple select * api * fontsize * code * chatid * fix build * fix * fix component * change name * log keys config * fix * delete unused * fix * chore: minio service class rewrite * chore: s3 plugin upload * feat: system global cache with multi node sync feature * feat: cache * chore: move images * docs: update & remove useless code * chore: resolve merge conflicts * chore: adjust the code * chore: adjust * deps: upgrade @fastgpt-sdk/plugin to 0.1.17 * perf(s3): s3 config * fix: cache syncKey refresh * fix: update @fastgpt-sdk/plugin to v0.1.18 removing mongo definition for fixing vitest * chore: adjust --------- Co-authored-by: Ctrlz <143257420+ctrlz526@users.noreply.github.com> Co-authored-by: heheer <heheer@sealos.io> Co-authored-by: Archer <545436317@qq.com> * perf: s3 api code * fix: toolbox empty when second open modal * feat: http tool set (#5599) * feat: http toolSet manual create front end * feat: http toolSet manual create i18n * feat: http toolSet manual create back end * feat: auth, as tool param, adapt mcp * fix: delete unused httpPlugin * fix: delete FlowNodeTypeEnum.httpPlugin * fix: AppTypeEnum include httpToolSet and httpPlugin * fix * delete console * fix * output schema * fix * fix bg * fix base url * fix --------- Co-authored-by: heheer <zhiyu44@qq.com> * feat: app count * perf: type check * feat: catch error * perf: plugin hot-swapping (#5688) * perf: plugin hot-swapping * chore: adjust code * perf: cite data auth * fix http toolset (#5689) * temp * fix http tool set * fix * template author hide * dynamic IO ui * fix: auth test * fix dynamic input & output (#5690) Co-authored-by: Archer <545436317@qq.com> * fix: dynamic output id * doc * feat: model permission (#5666) * feat(permission): model permission definition & api * chore: support update model's collaborators * feat: remove unauthedmodel when paste and import * fix: type error * fix: test setup global model list * fix: http tool api * chore: update fastgpt-sdk version * chore: remove useless code * chore: myModelList cache * perf: user who is not manager can not configure model permission (FE) * perf: model => Set * feat: getMyModels moved to opensource code; cache the myModelList * fix: type error * fix dynamic input reference select type (#5694) * remove unique index * read file usage * perf: connection error * fix: abort token count * fix: debug usage concat * fix: immer clone object * fix: immer clone object * perf: throw error when error chat * update audit i18n * fix: 修复识别pptx文件后,返回内容顺序错乱问题 (#5696) * fix: pptx sort error * fix prompt editor (#5695) * fix prompt editor * fix * fix: redis cache prefix (#5697) * fix: redis cache prefix * fix: cache * fix: get model collaborator by model.model * feat: hint for model per * rename bucket name * model ui * doc * doc --------- Co-authored-by: heheer <heheer@sealos.io> Co-authored-by: Finley Ge <32237950+FinleyGe@users.noreply.github.com> Co-authored-by: Ctrlz <143257420+ctrlz526@users.noreply.github.com> Co-authored-by: Zeng Qingwen <143274079+fishwww-ww@users.noreply.github.com> Co-authored-by: heheer <zhiyu44@qq.com> Co-authored-by: Deepturn <33342819+Deepturn@users.noreply.github.com>
95 lines
4.0 KiB
JSON
95 lines
4.0 KiB
JSON
{
|
|
"Hunyuan": "Tencent Hunyuan",
|
|
"aipoint_usage": "AI points",
|
|
"all": "All",
|
|
"api_key": "API key",
|
|
"avg_response_time": "Average call time (seconds)",
|
|
"avg_ttfb": "Average first word duration (seconds)",
|
|
"azure": "Azure",
|
|
"base_url": "Base url",
|
|
"batch_size": "Number of concurrent requests",
|
|
"channel_name": "Channel",
|
|
"channel_priority": "Priority",
|
|
"channel_priority_tip": "The higher the priority channel, the easier it is to be requested",
|
|
"channel_status": "state",
|
|
"channel_status_auto_disabled": "Automatically disable",
|
|
"channel_status_disabled": "Disabled",
|
|
"channel_status_enabled": "Enable",
|
|
"channel_status_unknown": "unknown",
|
|
"channel_type": "Protocol Type",
|
|
"clear_model": "Clear the model",
|
|
"confirm_delete_channel": "Confirm the deletion of the [{{name}}] channel?",
|
|
"copy_model_id_success": "Copyed model id",
|
|
"create_channel": "Added channels",
|
|
"dashboard_call_trend": "Model Call Trend",
|
|
"dashboard_channel": "Channel",
|
|
"dashboard_cost_trend": "Cost Consumption",
|
|
"dashboard_error_calls": "Error Calls",
|
|
"dashboard_input_tokens": "Input Tokens",
|
|
"dashboard_model": "Model",
|
|
"dashboard_no_data": "No data available",
|
|
"dashboard_output_tokens": "Output Tokens",
|
|
"dashboard_points": "points",
|
|
"dashboard_success_calls": "Success Calls",
|
|
"dashboard_token_trend": "Token Usage Trend",
|
|
"dashboard_token_usage": "Tokens",
|
|
"dashboard_total_calls": "Total Calls:",
|
|
"dashboard_total_cost": "Total Cost",
|
|
"dashboard_total_cost_label": "Total Cost:",
|
|
"dashboard_total_tokens": "Total Tokens",
|
|
"default_url": "Default address",
|
|
"detail": "Detail",
|
|
"duration": "Duration",
|
|
"edit": "edit",
|
|
"edit_channel": "Channel configuration",
|
|
"enable_channel": "Enable",
|
|
"forbid_channel": "Disabled",
|
|
"input": "Input",
|
|
"key_type": "API key format:",
|
|
"log": "Call log",
|
|
"log_detail": "Log details",
|
|
"log_request_id_search": "Search by requestId",
|
|
"log_status": "Status",
|
|
"mapping": "Model Mapping",
|
|
"mapping_tip": "A valid Json is required. \nThe model can be mapped when sending a request to the actual address. \nFor example:\n{\n \n \"gpt-4o\": \"gpt-4o-test\"\n\n}\n\nWhen FastGPT requests the gpt-4o model, the gpt-4o-test model is sent to the actual address, instead of gpt-4o.",
|
|
"maxToken_tip": "Model max_tokens parameter",
|
|
"max_rpm": "Max RPM (Requests Per Minute)",
|
|
"max_temperature_tip": "If the model temperature parameter is not filled in, it means that the model does not support the temperature parameter.",
|
|
"max_tpm": "Max TPM (Tokens Per Minute)",
|
|
"model": "Model",
|
|
"model_error_rate": "Error rate",
|
|
"model_error_request_times": "Number of failures",
|
|
"model_name": "Model name",
|
|
"model_request_times": "Request times",
|
|
"model_test": "Model testing",
|
|
"model_tokens": "Input/Output tokens",
|
|
"model_ttfb_time": "Response time of first word",
|
|
"monitoring": "Monitoring",
|
|
"output": "Output",
|
|
"request_at": "Request time",
|
|
"request_duration": "Request duration: {{duration}}s",
|
|
"retry_times": "Number of retry times",
|
|
"running_test": "In testing",
|
|
"search_model": "Search for models",
|
|
"select_channel": "Select a channel name",
|
|
"select_model": "Select a model",
|
|
"select_model_placeholder": "Select the model available under this channel",
|
|
"select_provider_placeholder": "Search protocol type",
|
|
"selected_model_empty": "Choose at least one model",
|
|
"start_test": "Batch test {{num}} models",
|
|
"test_failed": "There are {{num}} models that report errors",
|
|
"timespan_day": "Day",
|
|
"timespan_hour": "Hour",
|
|
"timespan_label": "Time Granularity",
|
|
"timespan_minute": "Minute",
|
|
"total_call_volume": "Request amount",
|
|
"use_in_eval": "Use in eval",
|
|
"view_chart": "Chart",
|
|
"view_table": "Table",
|
|
"vlm_model": "Vlm",
|
|
"vlm_model_tip": "Used to generate additional indexing of images in a document in the knowledge base",
|
|
"volunme_of_failed_calls": "Error amount",
|
|
"waiting_test": "Waiting for testing",
|
|
"model_permission_config_hint": "If no collaborators are added, all members are available by default"
|
|
}
|