mirror of
https://github.com/Yanyutin753/RefreshToV1Api.git
synced 2025-12-20 01:03:02 +08:00
Compare commits
13 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
1f0c5a31a0 | ||
|
|
ccc4ce3bc0 | ||
|
|
ba26e9c49a | ||
|
|
e82ffcc912 | ||
|
|
0701530bb3 | ||
|
|
2d421b0a71 | ||
|
|
b8b0a38f1e | ||
|
|
26e6af6317 | ||
|
|
b559ec151f | ||
|
|
25468bdf38 | ||
|
|
51f732abc6 | ||
|
|
971cdada64 | ||
|
|
38a10e8f2d |
2
.github/workflows/oaifree-docker-image.yml
vendored
2
.github/workflows/oaifree-docker-image.yml
vendored
@@ -42,6 +42,6 @@ jobs:
|
||||
push: true
|
||||
tags: |
|
||||
yangclivia/pandora-to-api:${{ steps.tag_name.outputs.tag }}
|
||||
yangclivia/pandora-to-api:0.7.9
|
||||
yangclivia/pandora-to-api:latest
|
||||
platforms: linux/amd64,linux/arm64
|
||||
build-args: TARGETPLATFORM=${{ matrix.platform }}
|
||||
|
||||
@@ -17,7 +17,7 @@
|
||||
|
||||
3. 支持直接把refresh_token作为请求key,方便接入one_api
|
||||
|
||||
4. 支持 gpt-4o 、gpt-4-s 、基本所有的GPTS
|
||||
4. 支持 gpt-4o 、gpt-4-s 、o1模型、基本所有的GPTS
|
||||
|
||||
* **oaiFree 的 backend-api 接口,无需打码**
|
||||
|
||||
@@ -37,6 +37,8 @@
|
||||
|
||||
- [x] 支持 代码解释器、联网、绘图
|
||||
|
||||
- [x] 支持 o1-mini 和 o1-preview
|
||||
|
||||
- [x] 支持 gpt-4-s
|
||||
|
||||
- [x] 支持 gpt-4o 和 gpt-4o-mini
|
||||
@@ -44,7 +46,7 @@
|
||||
- [x] 支持 gpt-3.5-turbo
|
||||
|
||||
- [x] 支持 gpts
|
||||
|
||||
|
||||
- [x] 支持 流式输出
|
||||
|
||||
- [x] 支持 非流式输出
|
||||
|
||||
@@ -12,10 +12,13 @@
|
||||
"gpt_3_5_new_name": "gpt-3.5-turbo",
|
||||
"gpt_4_o_new_name": "gpt-4-o,gpt-4o",
|
||||
"gpt_4_o_mini_new_name": "gpt-4o-mini",
|
||||
"o1_preview_new_name": "o1_preview",
|
||||
"o1_mini_new_name": "o1_mini",
|
||||
"need_delete_conversation_after_response": "true",
|
||||
"use_oaiusercontent_url": "false",
|
||||
"custom_arkose_url": "false",
|
||||
"arkose_urls": "",
|
||||
"upload_success_text": "`🤖 文件上传成功,搜索将不再提供额外信息!`\n",
|
||||
"dalle_prompt_prefix": "请严格根据我的以下要求完成绘图任务,如果我没有发出指定的绘画指令,则绘制出我发出的文字对应的图片:",
|
||||
"bot_mode": {
|
||||
"enabled": "false",
|
||||
|
||||
@@ -2,7 +2,7 @@ version: '3'
|
||||
|
||||
services:
|
||||
backend-to-api:
|
||||
image: yangclivia/pandora-to-api:0.7.9
|
||||
image: yangclivia/pandora-to-api
|
||||
restart: always
|
||||
ports:
|
||||
- "50011:33333"
|
||||
|
||||
143
main.py
143
main.py
@@ -5,18 +5,17 @@ import json
|
||||
import logging
|
||||
import mimetypes
|
||||
import os
|
||||
import requests
|
||||
import uuid
|
||||
from datetime import datetime
|
||||
from io import BytesIO
|
||||
from logging.handlers import TimedRotatingFileHandler
|
||||
from queue import Queue
|
||||
from urllib.parse import urlparse
|
||||
|
||||
import requests
|
||||
from fake_useragent import UserAgent
|
||||
from flask import Flask, request, jsonify, Response, send_from_directory
|
||||
from flask_apscheduler import APScheduler
|
||||
from flask_cors import CORS, cross_origin
|
||||
from io import BytesIO
|
||||
from logging.handlers import TimedRotatingFileHandler
|
||||
from queue import Queue
|
||||
from urllib.parse import urlparse
|
||||
|
||||
|
||||
# 读取配置文件
|
||||
@@ -43,6 +42,9 @@ GPT_4_MOBILE_NEW_NAMES = CONFIG.get('gpt_4_mobile_new_name', 'gpt-4-mobile').spl
|
||||
GPT_3_5_NEW_NAMES = CONFIG.get('gpt_3_5_new_name', 'gpt-3.5-turbo').split(',')
|
||||
GPT_4_O_NEW_NAMES = CONFIG.get('gpt_4_o_new_name', 'gpt-4o').split(',')
|
||||
GPT_4_O_MINI_NEW_NAMES = CONFIG.get('gpt_4_o_mini_new_name', 'gpt-4o-mini').split(',')
|
||||
O1_PREVIEW_NEW_NAMES = CONFIG.get('o1_preview_new_name', 'o1-preview').split(',')
|
||||
O1_MINI_NEW_NAMES = CONFIG.get('o1_mini_new_name', 'o1-mini').split(',')
|
||||
UPLOAD_SUCCESS_TEXT = CONFIG.get('upload_success_text', "`🤖 文件上传成功,搜索将不再提供额外信息!`\n")
|
||||
|
||||
BOT_MODE = CONFIG.get('bot_mode', {})
|
||||
BOT_MODE_ENABLED = BOT_MODE.get('enabled', 'false').lower() == 'true'
|
||||
@@ -325,9 +327,9 @@ scheduler.start()
|
||||
# PANDORA_UPLOAD_URL = 'files.pandoranext.com'
|
||||
|
||||
|
||||
VERSION = '0.7.9.5'
|
||||
VERSION = '0.8.1'
|
||||
# VERSION = 'test'
|
||||
UPDATE_INFO = '✨ 支持最新的gpt-4o-mini 模型'
|
||||
UPDATE_INFO = '👀 支持输出o1思考过程'
|
||||
# UPDATE_INFO = '【仅供临时测试使用】 '
|
||||
|
||||
with app.app_context():
|
||||
@@ -448,6 +450,16 @@ with app.app_context():
|
||||
"name": name.strip(),
|
||||
"ori_name": "gpt-4o-mini"
|
||||
})
|
||||
for name in O1_PREVIEW_NEW_NAMES:
|
||||
gpts_configurations.append({
|
||||
"name": name.strip(),
|
||||
"ori_name": "o1-preview"
|
||||
})
|
||||
for name in O1_MINI_NEW_NAMES:
|
||||
gpts_configurations.append({
|
||||
"name": name.strip(),
|
||||
"ori_name": "o1-mini"
|
||||
})
|
||||
logger.info(f"GPTS 配置信息")
|
||||
|
||||
# 加载配置并添加到全局列表
|
||||
@@ -925,6 +937,63 @@ def send_text_prompt_and_get_response(messages, api_key, account_id, stream, mod
|
||||
"force_paragen_model_slug": "",
|
||||
"force_rate_limit": False
|
||||
}
|
||||
elif ori_model_name == 'o1-preview':
|
||||
payload = {
|
||||
"action": "next",
|
||||
"messages": formatted_messages,
|
||||
"parent_message_id": str(uuid.uuid4()),
|
||||
"model": "o1-preview",
|
||||
"timezone_offset_min": -480,
|
||||
"suggestions": [
|
||||
"What are 5 creative things I could do with my kids' art? I don't want to throw them away, "
|
||||
"but it's also so much clutter.",
|
||||
"I want to cheer up my friend who's having a rough day. Can you suggest a couple short and sweet "
|
||||
"text messages to go with a kitten gif?",
|
||||
"Come up with 5 concepts for a retro-style arcade game.",
|
||||
"I have a photoshoot tomorrow. Can you recommend me some colors and outfit options that will look "
|
||||
"good on camera?"
|
||||
],
|
||||
"variant_purpose": "comparison_implicit",
|
||||
"history_and_training_disabled": False,
|
||||
"conversation_mode": {
|
||||
"kind": "primary_assistant"
|
||||
},
|
||||
"force_paragen": False,
|
||||
"force_paragen_model_slug": "",
|
||||
"force_nulligen": False,
|
||||
"force_rate_limit": False,
|
||||
"reset_rate_limits": False,
|
||||
"force_use_sse": True,
|
||||
}
|
||||
elif ori_model_name == 'o1-mini':
|
||||
payload = {
|
||||
"action": "next",
|
||||
"messages": formatted_messages,
|
||||
"parent_message_id": str(uuid.uuid4()),
|
||||
"model": "o1-mini",
|
||||
"timezone_offset_min": -480,
|
||||
"suggestions": [
|
||||
"What are 5 creative things I could do with my kids' art? I don't want to throw them away, "
|
||||
"but it's also so much clutter.",
|
||||
"I want to cheer up my friend who's having a rough day. Can you suggest a couple short and sweet "
|
||||
"text messages to go with a kitten gif?",
|
||||
"Come up with 5 concepts for a retro-style arcade game.",
|
||||
"I have a photoshoot tomorrow. Can you recommend me some colors and outfit options that will look "
|
||||
"good on camera?"
|
||||
],
|
||||
"variant_purpose": "comparison_implicit",
|
||||
"history_and_training_disabled": False,
|
||||
"conversation_mode": {
|
||||
"kind": "primary_assistant"
|
||||
},
|
||||
"force_paragen": False,
|
||||
"force_paragen_model_slug": "",
|
||||
"force_nulligen": False,
|
||||
"force_rate_limit": False,
|
||||
"reset_rate_limits": False,
|
||||
"force_use_sse": True,
|
||||
}
|
||||
|
||||
elif 'gpt-4-gizmo-' in model:
|
||||
payload = generate_gpts_payload(model, formatted_messages)
|
||||
if not payload:
|
||||
@@ -1231,9 +1300,11 @@ def data_fetcher(upstream_response, data_queue, stop_event, last_data_time, api_
|
||||
while 'data:' in buffer and '\n\n' in buffer:
|
||||
end_index = buffer.index('\n\n') + 2
|
||||
complete_data, buffer = buffer[:end_index], buffer[end_index:]
|
||||
# 解析 data 块
|
||||
try:
|
||||
data_json = json.loads(complete_data.replace('data: ', ''))
|
||||
data_content = complete_data.replace('data: ', '').strip()
|
||||
if not data_content:
|
||||
continue
|
||||
data_json = json.loads(data_content)
|
||||
# print(f"data_json: {data_json}")
|
||||
message = data_json.get("message", {})
|
||||
|
||||
@@ -1373,9 +1444,10 @@ def data_fetcher(upstream_response, data_queue, stop_event, last_data_time, api_
|
||||
# 只获取新的 parts
|
||||
parts = content.get("parts", [])
|
||||
full_text = ''.join(parts)
|
||||
if full_text == "![":
|
||||
last_full_text = "!"
|
||||
new_text = full_text[len(last_full_text):]
|
||||
if full_text != '':
|
||||
last_full_text = full_text # 更新完整文本以备下次比较
|
||||
last_full_text = full_text
|
||||
if "\u3010" in new_text and not citation_accumulating:
|
||||
citation_accumulating = True
|
||||
citation_buffer = citation_buffer + new_text
|
||||
@@ -1492,13 +1564,13 @@ def data_fetcher(upstream_response, data_queue, stop_event, last_data_time, api_
|
||||
last_full_code_result = full_code_result
|
||||
|
||||
# 其余Action执行输出特殊处理
|
||||
if role == "tool" and name != "python" and name != "dalle.text2im" and last_content_type != "execution_output" and content_type != None:
|
||||
new_text = ""
|
||||
if last_content_type == "code":
|
||||
if BOT_MODE_ENABLED and BOT_MODE_ENABLED_CODE_BLOCK_OUTPUT == False:
|
||||
new_text = ""
|
||||
else:
|
||||
new_text = "\n```\n" + new_text
|
||||
# if role == "tool" and name != "python" and name != "dalle.text2im" and last_content_type != "execution_output" and content_type != None:
|
||||
# new_text = ""
|
||||
# if last_content_type == "code":
|
||||
# if BOT_MODE_ENABLED and BOT_MODE_ENABLED_CODE_BLOCK_OUTPUT == False:
|
||||
# new_text = ""
|
||||
# else:
|
||||
# new_text = "\n```\n" + new_text
|
||||
|
||||
# 检查 new_text 中是否包含 <<ImageDisplayed>>
|
||||
if "<<ImageDisplayed>>" in last_full_code_result:
|
||||
@@ -1545,6 +1617,10 @@ def data_fetcher(upstream_response, data_queue, stop_event, last_data_time, api_
|
||||
execution_output_image_id_buffer = image_file_id
|
||||
|
||||
# 从 new_text 中移除 <<ImageDisplayed>>
|
||||
new_text = new_text.replace(
|
||||
"All the files uploaded by the user have been fully loaded. Searching won't provide "
|
||||
"additional information.",
|
||||
UPLOAD_SUCCESS_TEXT)
|
||||
new_text = new_text.replace("<<ImageDisplayed>>", "图片生成中,请稍后\n")
|
||||
|
||||
# print(f"收到数据: {data_json}")
|
||||
@@ -1795,14 +1871,22 @@ import time
|
||||
def chat_completions():
|
||||
logger.info(f"New Request")
|
||||
proxy_api_prefix = getPROXY_API_PREFIX(lock)
|
||||
|
||||
if proxy_api_prefix == None:
|
||||
return jsonify({"error": "PROXY_API_PREFIX is not accessible"}), 401
|
||||
data = request.json
|
||||
messages = data.get('messages')
|
||||
model = data.get('model')
|
||||
model = data.get('model', "gpt-3.5-turbo")
|
||||
ori_model_name = model
|
||||
accessible_model_list = get_accessible_model_list()
|
||||
if model not in accessible_model_list and not 'gpt-4-gizmo-' in model:
|
||||
return jsonify({"error": "model is not accessible"}), 401
|
||||
model_config = find_model_config(model)
|
||||
if model_config:
|
||||
ori_model_name = model_config.get('ori_name', model)
|
||||
if "o1-" in ori_model_name:
|
||||
# 使用列表推导式过滤系统角色
|
||||
messages = [message for message in messages if message["role"] in ["user", "assistant"]]
|
||||
|
||||
stream = data.get('stream', False)
|
||||
|
||||
@@ -1971,18 +2055,25 @@ def images_generations():
|
||||
data = request.json
|
||||
logger.debug(f"data: {data}")
|
||||
api_key = None
|
||||
# messages = data.get('messages')
|
||||
model = data.get('model')
|
||||
model = data.get('model', "gpt-3.5-turbo")
|
||||
ori_model_name = model
|
||||
accessible_model_list = get_accessible_model_list()
|
||||
if model not in accessible_model_list and not 'gpt-4-gizmo-' in model:
|
||||
return jsonify({"error": "model is not accessible"}), 401
|
||||
model_config = find_model_config(model)
|
||||
if model_config:
|
||||
ori_model_name = model_config.get('ori_name', model)
|
||||
if "o1-" in ori_model_name:
|
||||
# 使用列表推导式过滤系统角色
|
||||
messages = [message for message in messages if message["role"] in ["user", "assistant"]]
|
||||
# 获取请求中的response_format参数,默认为"url"
|
||||
response_format = data.get('response_format', 'url')
|
||||
# 获取请求中的size参数,默认为"1024x1024"
|
||||
response_size = data.get('size', '1024x1024')
|
||||
|
||||
prompt = data.get('prompt', '')
|
||||
|
||||
prompt = DALLE_PROMPT_PREFIX + prompt
|
||||
|
||||
# 获取请求中的response_format参数,默认为"url"
|
||||
response_format = data.get('response_format', 'url')
|
||||
prompt = DALLE_PROMPT_PREFIX + '\nprompt:' + prompt + '\nsize:' + response_size
|
||||
|
||||
# stream = data.get('stream', False)
|
||||
|
||||
|
||||
Reference in New Issue
Block a user