fix: empty response

This commit is contained in:
Vincent Yang
2024-04-21 01:56:21 -04:00
parent 6b72d71ba0
commit b4c6227075
3 changed files with 170 additions and 63 deletions

208
app.js
View File

@@ -2,14 +2,20 @@ const express = require("express");
const bodyParser = require("body-parser"); const bodyParser = require("body-parser");
const axios = require("axios"); const axios = require("axios");
const https = require("https"); const https = require("https");
const { randomUUID } = require("crypto"); const { encode } = require("gpt-3-encoder");
const { randomUUID, randomInt, createHash } = require("crypto");
const { config } = require("dotenv");
config();
// Constants for the server and API configuration // Constants for the server and API configuration
const port = 3040; const port = process.env.SERVER_PORT || 3040;
const baseUrl = "https://chat.openai.com"; const baseUrl = "https://chat.openai.com";
const apiUrl = `${baseUrl}/backend-api/conversation`; const apiUrl = `${baseUrl}/backend-api/conversation`;
const refreshInterval = 60000; // Interval to refresh token in ms const refreshInterval = 60000; // Interval to refresh token in ms
const errorWait = 120000; // Wait time in ms after an error const errorWait = 120000; // Wait time in ms after an error
const newSessionRetries = parseInt(process.env.NEW_SESSION_RETRIES) || 5;
const userAgent = process.env.USER_AGENT || "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/123.0.0.0 Safari/537.36";
// Initialize global variables to store the session token and device ID // Initialize global variables to store the session token and device ID
let token; let token;
@@ -19,8 +25,7 @@ let oaiDeviceId;
const wait = (ms) => new Promise((resolve) => setTimeout(resolve, ms)); const wait = (ms) => new Promise((resolve) => setTimeout(resolve, ms));
function GenerateCompletionId(prefix = "cmpl-") { function GenerateCompletionId(prefix = "cmpl-") {
const characters = const characters = "abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789";
"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789";
const length = 28; const length = 28;
for (let i = 0; i < length; i++) { for (let i = 0; i < length; i++) {
@@ -70,38 +75,69 @@ const axiosInstance = axios.create({
origin: baseUrl, origin: baseUrl,
pragma: "no-cache", pragma: "no-cache",
referer: baseUrl, referer: baseUrl,
"sec-ch-ua": "sec-ch-ua": '"Google Chrome";v="123", "Not:A-Brand";v="8", "Chromium";v="123"',
'"Google Chrome";v="123", "Not:A-Brand";v="8", "Chromium";v="123"',
"sec-ch-ua-mobile": "?0", "sec-ch-ua-mobile": "?0",
"sec-ch-ua-platform": '"Windows"', "sec-ch-ua-platform": '"Windows"',
"sec-fetch-dest": "empty", "sec-fetch-dest": "empty",
"sec-fetch-mode": "cors", "sec-fetch-mode": "cors",
"sec-fetch-site": "same-origin", "sec-fetch-site": "same-origin",
"user-agent": "user-agent": userAgent,
"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/123.0.0.0 Safari/537.36",
}, },
}); });
// Function to get a new session ID and token from the OpenAI API function generateProofToken(seed, diff, userAgent) {
async function getNewSessionId() { const cores = [8, 12, 16, 24];
let newDeviceId = randomUUID(); const screens = [3000, 4000, 6000];
const response = await axiosInstance.post( const core = cores[randomInt(0, cores.length)];
`${baseUrl}/backend-anon/sentinel/chat-requirements`, const screen = screens[randomInt(0, screens.length)];
{}, const now = new Date(Date.now() - 8 * 3600 * 1000);
{ const parseTime = now.toUTCString().replace("GMT", "GMT-0500 (Eastern Time)");
headers: { "oai-device-id": newDeviceId }, const config = [core + screen, parseTime, 4294705152, 0, userAgent];
const diffLen = diff.length / 2;
for (let i = 0; i < 100000; i++) {
config[3] = i;
const jsonData = JSON.stringify(config);
const base = Buffer.from(jsonData).toString("base64");
const hashValue = createHash("sha3-512")
.update(seed + base)
.digest();
if (hashValue.toString("hex").substring(0, diffLen) <= diff) {
const result = "gAAAAAB" + base;
return result;
} }
); }
console.log( const fallbackBase = Buffer.from(`"${seed}"`).toString("base64");
`System: Successfully refreshed session ID and token. ${ return "gAAAAABwQ8Lk5FbGpA2NcR9dShT6gYjU7VxZ4D" + fallbackBase;
!token ? "(Now it's ready to process requests)" : "" }
}`
);
oaiDeviceId = newDeviceId;
token = response.data.token;
// console.log("New Token:", token); // Function to get a new session ID and token from the OpenAI API
// console.log("New Device ID:", oaiDeviceId); async function getNewSession(retries = 0) {
let newDeviceId = randomUUID();
try {
const response = await axiosInstance.post(
`${baseUrl}/backend-anon/sentinel/chat-requirements`,
{},
{
headers: { "oai-device-id": newDeviceId },
}
);
let session = response.data;
session.deviceId = newDeviceId;
console.log(
`System: Successfully refreshed session ID and token. ${
!token ? "(Now it's ready to process requests)" : ""
}`
);
oaiDeviceId = newDeviceId;
token = session.token;
return session;
} catch (error) {
await wait(500);
return retries < newSessionRetries ? getNewSession(retries + 1) : null;
}
} }
// Middleware to enable CORS and handle pre-flight requests // Middleware to enable CORS and handle pre-flight requests
@@ -120,10 +156,28 @@ async function handleChatCompletion(req, res) {
console.log( console.log(
"Request:", "Request:",
`${req.method} ${req.originalUrl}`, `${req.method} ${req.originalUrl}`,
`${req.body?.messages?.length || 0} messages`, `${req.body?.messages?.length ?? 0} messages`,
req.body.stream ? "(stream-enabled)" : "(stream-disabled)" req.body.stream ? "(stream-enabled)" : "(stream-disabled)"
); );
try { try {
let session = await getNewSession();
if (!session) {
res.write(
JSON.stringify({
status: false,
error: {
message: `Error getting a new session, please try again later, if the issue persists, please open an issue on the GitHub repository.`,
type: "invalid_request_error",
},
})
);
return res.end();
}
let proofToken = generateProofToken(
session.proofofwork.seed,
session.proofofwork.difficulty,
userAgent
);
const body = { const body = {
action: "next", action: "next",
messages: req.body.messages.map((message) => ({ messages: req.body.messages.map((message) => ({
@@ -138,12 +192,18 @@ async function handleChatCompletion(req, res) {
conversation_mode: { kind: "primary_assistant" }, conversation_mode: { kind: "primary_assistant" },
websocket_request_id: randomUUID(), websocket_request_id: randomUUID(),
}; };
let promptTokens = 0;
let completionTokens = 0;
for (let message of req.body.messages) {
promptTokens += encode(message.content).length;
}
const response = await axiosInstance.post(apiUrl, body, { const response = await axiosInstance.post(apiUrl, body, {
responseType: "stream", responseType: "stream",
headers: { headers: {
"oai-device-id": oaiDeviceId, "oai-device-id": session.deviceId,
"openai-sentinel-chat-requirements-token": token, "openai-sentinel-chat-requirements-token": session.token,
"openai-sentinel-proof-token": proofToken,
}, },
}); });
@@ -158,15 +218,25 @@ async function handleChatCompletion(req, res) {
let fullContent = ""; let fullContent = "";
let requestId = GenerateCompletionId("chatcmpl-"); let requestId = GenerateCompletionId("chatcmpl-");
let created = Date.now(); let created = Math.floor(Date.now() / 1000); // Unix timestamp in seconds
let finish_reason = null;
let error;
for await (const message of StreamCompletion(response.data)) { for await (const message of StreamCompletion(response.data)) {
// Skip heartbeat detection // Skip heartbeat detection
if (message.match(/^\d{4}-\d{2}-\d{2} \d{2}:\d{2}:\d{2}.\d{6}$/)) continue; if (message.match(/^\d{4}-\d{2}-\d{2} \d{2}:\d{2}:\d{2}.\d{6}$/))
continue;
const parsed = JSON.parse(message); const parsed = JSON.parse(message);
let content = parsed?.message?.content?.parts[0] || ""; if (parsed.error) {
error = `Error message from OpenAI: ${parsed.error}`;
finish_reason = "stop";
break;
}
let content = parsed?.message?.content?.parts[0] ?? "";
let status = parsed?.message?.status ?? "";
for (let message of req.body.messages) { for (let message of req.body.messages) {
if (message.content === content) { if (message.content === content) {
@@ -175,8 +245,32 @@ async function handleChatCompletion(req, res) {
} }
} }
switch (status) {
case "in_progress":
finish_reason = null;
break;
case "finished_successfully":
let finish_reason_data =
parsed?.message?.metadata?.finish_details?.type ?? null;
switch (finish_reason_data) {
case "max_tokens":
finish_reason = "length";
break;
case "stop":
default:
finish_reason = "stop";
}
break;
default:
finish_reason = null;
}
if (content === "") continue; if (content === "") continue;
let completionChunk = content.replace(fullContent, "");
completionTokens += encode(completionChunk).length;
if (req.body.stream) { if (req.body.stream) {
let response = { let response = {
id: requestId, id: requestId,
@@ -186,10 +280,10 @@ async function handleChatCompletion(req, res) {
choices: [ choices: [
{ {
delta: { delta: {
content: content.replace(fullContent, ""), content: completionChunk,
}, },
index: 0, index: 0,
finish_reason: null, finish_reason: finish_reason,
}, },
], ],
}; };
@@ -210,10 +304,10 @@ async function handleChatCompletion(req, res) {
choices: [ choices: [
{ {
delta: { delta: {
content: "", content: error ?? "",
}, },
index: 0, index: 0,
finish_reason: "stop", finish_reason: finish_reason,
}, },
], ],
})}\n\n` })}\n\n`
@@ -227,18 +321,18 @@ async function handleChatCompletion(req, res) {
object: "chat.completion", object: "chat.completion",
choices: [ choices: [
{ {
finish_reason: "stop", finish_reason: finish_reason,
index: 0, index: 0,
message: { message: {
content: fullContent, content: error ?? fullContent,
role: "assistant", role: "assistant",
}, },
}, },
], ],
usage: { usage: {
prompt_tokens: 0, prompt_tokens: promptTokens,
completion_tokens: 0, completion_tokens: completionTokens,
total_tokens: 0, total_tokens: promptTokens + completionTokens,
}, },
}) })
); );
@@ -246,15 +340,13 @@ async function handleChatCompletion(req, res) {
res.end(); res.end();
} catch (error) { } catch (error) {
// console.log('Error:', error.response?.data ?? error.message);
if (!res.headersSent) res.setHeader("Content-Type", "application/json"); if (!res.headersSent) res.setHeader("Content-Type", "application/json");
// console.error('Error handling chat completion:', error);
res.write( res.write(
JSON.stringify({ JSON.stringify({
status: false, status: false,
error: { error: {
message: message:
"An error happened, please make sure your request is SFW, or use a jailbreak to bypass the filter.", "An error occurred. Please try again. Additionally, ensure that your request complies with OpenAI's policy.",
type: "invalid_request_error", type: "invalid_request_error",
}, },
}) })
@@ -276,40 +368,34 @@ app.use((req, res) =>
res.status(404).send({ res.status(404).send({
status: false, status: false,
error: { error: {
message: `The requested endpoint was not found. please make sure to use "http://localhost:3040/v1" as the base URL.`, message: `The requested endpoint (${req.method.toLocaleUpperCase()} ${req.path}) was not found. please make sure what are you doing now.`,
type: "invalid_request_error", type: "invalid_request_error",
}, },
}) })
); );
// Start the server and the session ID refresh loop // Start the server and the session ID refresh loop
app.listen(port, () => { app.listen(port, async () => {
console.log(`💡 Server is running at http://localhost:${port}`); console.log(`💡 Server is running at http://localhost:${port}`);
console.log(); console.log();
console.log(`🔗 Base URL: http://localhost:${port}/v1`); console.log(`🔗 Local Base URL: http://localhost:${port}/v1`);
console.log( console.log(`🔗 Local Endpoint: http://localhost:${port}/v1/chat/completions`);
`🔗 ChatCompletion Endpoint: http://localhost:${port}/v1/chat/completions`
);
console.log(); console.log();
console.log("📝 Original TS Source By: Pawan.Krd"); console.log("📝 Original TS Source By: Pawan.Krd");
console.log("📝 Modified Into JavaScript By: Adam"); console.log("📝 Modified By: Vincent");
console.log(); console.log();
setTimeout(async () => { setTimeout(async () => {
while (true) { while (true) {
try { try {
await getNewSessionId(); await getNewSession();
await wait(refreshInterval); await wait(refreshInterval);
} catch (error) { } catch (error) {
console.error("Error refreshing session ID, retrying in 1 minute..."); console.error("Error refreshing session ID, retrying in 2 minute...");
console.error( console.error("If this error persists, your country may not be supported yet.");
"If this error persists, your country may not be supported yet." console.error("If your country was the issue, please consider using a U.S. VPN.");
);
console.error(
"If your country was the issue, please consider using a U.S. VPN."
);
await wait(errorWait); await wait(errorWait);
} }
} }
}, 0); }, 0);
}); });

21
package-lock.json generated
View File

@@ -6,7 +6,10 @@
"": { "": {
"dependencies": { "dependencies": {
"axios": "^1.6.8", "axios": "^1.6.8",
"express": "^4.19.2" "body-parser": "^1.20.2",
"dotenv": "^16.4.5",
"express": "^4.19.2",
"gpt-3-encoder": "^1.1.4"
} }
}, },
"node_modules/accepts": { "node_modules/accepts": {
@@ -182,6 +185,17 @@
"npm": "1.2.8000 || >= 1.4.16" "npm": "1.2.8000 || >= 1.4.16"
} }
}, },
"node_modules/dotenv": {
"version": "16.4.5",
"resolved": "https://registry.npmjs.org/dotenv/-/dotenv-16.4.5.tgz",
"integrity": "sha512-ZmdL2rui+eB2YwhsWzjInR8LldtZHGDoQ1ugH85ppHKwpUHL7j7rN0Ti9NCnGiQbhaZ11FpR+7ao1dNsmduNUg==",
"engines": {
"node": ">=12"
},
"funding": {
"url": "https://dotenvx.com"
}
},
"node_modules/ee-first": { "node_modules/ee-first": {
"version": "1.1.1", "version": "1.1.1",
"resolved": "https://registry.npmjs.org/ee-first/-/ee-first-1.1.1.tgz", "resolved": "https://registry.npmjs.org/ee-first/-/ee-first-1.1.1.tgz",
@@ -370,6 +384,11 @@
"url": "https://github.com/sponsors/ljharb" "url": "https://github.com/sponsors/ljharb"
} }
}, },
"node_modules/gpt-3-encoder": {
"version": "1.1.4",
"resolved": "https://registry.npmjs.org/gpt-3-encoder/-/gpt-3-encoder-1.1.4.tgz",
"integrity": "sha512-fSQRePV+HUAhCn7+7HL7lNIXNm6eaFWFbNLOOGtmSJ0qJycyQvj60OvRlH7mee8xAMjBDNRdMXlMwjAbMTDjkg=="
},
"node_modules/has-property-descriptors": { "node_modules/has-property-descriptors": {
"version": "1.0.2", "version": "1.0.2",
"resolved": "https://registry.npmjs.org/has-property-descriptors/-/has-property-descriptors-1.0.2.tgz", "resolved": "https://registry.npmjs.org/has-property-descriptors/-/has-property-descriptors-1.0.2.tgz",

View File

@@ -2,6 +2,8 @@
"dependencies": { "dependencies": {
"axios": "^1.6.8", "axios": "^1.6.8",
"body-parser": "^1.20.2", "body-parser": "^1.20.2",
"express": "^4.19.2" "dotenv": "^16.4.5",
"express": "^4.19.2",
"gpt-3-encoder": "^1.1.4"
} }
} }