Files
FastGPT/packages/web/i18n/en/account_model.json
Archer 101a6e9516 Feat: ai proxy monitor (#4985)
* Aiproxy ModelBoard (#4983)

* Aiproxy ModelBoard

* Add components LineChartComponent and Make some revisions

* perf: ai proxy dashboard

* doc

* remove invalid i18n

* remove invalid i18n

---------

Co-authored-by: Zhuangzai fa <143257420+ctrlz526@users.noreply.github.com>
2025-06-10 01:46:10 +08:00

67 lines
2.8 KiB
JSON

{
"Hunyuan": "Tencent Hunyuan",
"aipoint_usage": "AI points",
"all": "All",
"api_key": "API key",
"azure": "Azure",
"base_url": "Base url",
"channel_name": "Channel",
"channel_priority": "Priority",
"channel_priority_tip": "The higher the priority channel, the easier it is to be requested",
"channel_status": "state",
"channel_status_auto_disabled": "Automatically disable",
"channel_status_disabled": "Disabled",
"channel_status_enabled": "Enable",
"channel_status_unknown": "unknown",
"channel_type": "Manufacturer",
"clear_model": "Clear the model",
"confirm_delete_channel": "Confirm the deletion of the [{{name}}] channel?",
"copy_model_id_success": "Copyed model id",
"create_channel": "Added channels",
"dashboard_error_calls": "Error Calls",
"dashboard_model": "Model",
"dashboard_points": "points",
"dashboard_token_usage": "Tokens",
"default_url": "Default address",
"detail": "Detail",
"duration": "Duration",
"edit": "edit",
"edit_channel": "Channel configuration",
"enable_channel": "Enable",
"forbid_channel": "Disabled",
"input": "Input",
"key_type": "API key format:",
"log": "Call log",
"log_detail": "Log details",
"log_request_id_search": "Search by requestId",
"log_status": "Status",
"mapping": "Model Mapping",
"mapping_tip": "A valid Json is required. \nThe model can be mapped when sending a request to the actual address. \nFor example:\n{\n \n \"gpt-4o\": \"gpt-4o-test\"\n\n}\n\nWhen FastGPT requests the gpt-4o model, the gpt-4o-test model is sent to the actual address, instead of gpt-4o.",
"maxToken_tip": "Model max_tokens parameter",
"max_temperature_tip": "If the model temperature parameter is not filled in, it means that the model does not support the temperature parameter.",
"model": "Model",
"model_error_rate": "Error rate",
"model_error_request_times": "Number of failures",
"model_name": "Model name",
"model_request_times": "Request times",
"model_test": "Model testing",
"model_tokens": "Input/Output tokens",
"monitoring": "Monitoring",
"output": "Output",
"request_at": "Request time",
"request_duration": "Request duration: {{duration}}s",
"retry_times": "Number of retry times",
"running_test": "In testing",
"search_model": "Search for models",
"select_channel": "Select a channel name",
"select_model": "Select a model",
"select_model_placeholder": "Select the model available under this channel",
"select_provider_placeholder": "Search for manufacturers",
"selected_model_empty": "Choose at least one model",
"start_test": "Batch test {{num}} models",
"test_failed": "There are {{num}} models that report errors",
"vlm_model": "Vlm",
"vlm_model_tip": "Used to generate additional indexing of images in a document in the knowledge base",
"waiting_test": "Waiting for testing"
}