mirror of
https://github.com/SillyTavern/SillyTavern.git
synced 2025-06-05 21:59:27 +02:00
Merge pull request #3852 from subzero5544/xAI-grok-reverse-proxy-testing
Adding reverse proxy support to xai chat completion
This commit is contained in:
@ -2762,7 +2762,7 @@
|
|||||||
<option value="xai">xAI (Grok)</option>
|
<option value="xai">xAI (Grok)</option>
|
||||||
</optgroup>
|
</optgroup>
|
||||||
</select>
|
</select>
|
||||||
<div class="inline-drawer wide100p" data-source="openai,claude,mistralai,makersuite,deepseek">
|
<div class="inline-drawer wide100p" data-source="openai,claude,mistralai,makersuite,deepseek,xai">
|
||||||
<div class="inline-drawer-toggle inline-drawer-header">
|
<div class="inline-drawer-toggle inline-drawer-header">
|
||||||
<b data-i18n="Reverse Proxy">Reverse Proxy</b>
|
<b data-i18n="Reverse Proxy">Reverse Proxy</b>
|
||||||
<div class="fa-solid fa-circle-chevron-down inline-drawer-icon down"></div>
|
<div class="fa-solid fa-circle-chevron-down inline-drawer-icon down"></div>
|
||||||
@ -2825,7 +2825,7 @@
|
|||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
<div id="ReverseProxyWarningMessage" data-source="openai,claude,mistralai,makersuite,deepseek">
|
<div id="ReverseProxyWarningMessage" data-source="openai,claude,mistralai,makersuite,deepseek,xai">
|
||||||
<div class="reverse_proxy_warning">
|
<div class="reverse_proxy_warning">
|
||||||
<b>
|
<b>
|
||||||
<div data-i18n="Using a proxy that you're not running yourself is a risk to your data privacy.">
|
<div data-i18n="Using a proxy that you're not running yourself is a risk to your data privacy.">
|
||||||
|
@ -2039,7 +2039,7 @@ async function sendOpenAIRequest(type, messages, signal) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Proxy is only supported for Claude, OpenAI, Mistral, and Google MakerSuite
|
// Proxy is only supported for Claude, OpenAI, Mistral, and Google MakerSuite
|
||||||
if (oai_settings.reverse_proxy && [chat_completion_sources.CLAUDE, chat_completion_sources.OPENAI, chat_completion_sources.MISTRALAI, chat_completion_sources.MAKERSUITE, chat_completion_sources.DEEPSEEK].includes(oai_settings.chat_completion_source)) {
|
if (oai_settings.reverse_proxy && [chat_completion_sources.CLAUDE, chat_completion_sources.OPENAI, chat_completion_sources.MISTRALAI, chat_completion_sources.MAKERSUITE, chat_completion_sources.DEEPSEEK, chat_completion_sources.XAI].includes(oai_settings.chat_completion_source)) {
|
||||||
await validateReverseProxy();
|
await validateReverseProxy();
|
||||||
generate_data['reverse_proxy'] = oai_settings.reverse_proxy;
|
generate_data['reverse_proxy'] = oai_settings.reverse_proxy;
|
||||||
generate_data['proxy_password'] = oai_settings.proxy_password;
|
generate_data['proxy_password'] = oai_settings.proxy_password;
|
||||||
@ -3545,7 +3545,7 @@ async function getStatusOpen() {
|
|||||||
chat_completion_source: oai_settings.chat_completion_source,
|
chat_completion_source: oai_settings.chat_completion_source,
|
||||||
};
|
};
|
||||||
|
|
||||||
if (oai_settings.reverse_proxy && [chat_completion_sources.CLAUDE, chat_completion_sources.OPENAI, chat_completion_sources.MISTRALAI, chat_completion_sources.MAKERSUITE, chat_completion_sources.DEEPSEEK].includes(oai_settings.chat_completion_source)) {
|
if (oai_settings.reverse_proxy && [chat_completion_sources.CLAUDE, chat_completion_sources.OPENAI, chat_completion_sources.MISTRALAI, chat_completion_sources.MAKERSUITE, chat_completion_sources.DEEPSEEK, chat_completion_sources.XAI].includes(oai_settings.chat_completion_source)) {
|
||||||
await validateReverseProxy();
|
await validateReverseProxy();
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -4943,7 +4943,7 @@ async function onConnectButtonClick(e) {
|
|||||||
await writeSecret(SECRET_KEYS.XAI, api_key_xai);
|
await writeSecret(SECRET_KEYS.XAI, api_key_xai);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!secret_state[SECRET_KEYS.XAI]) {
|
if (!secret_state[SECRET_KEYS.XAI] && !oai_settings.reverse_proxy) {
|
||||||
console.log('No secret key saved for XAI');
|
console.log('No secret key saved for XAI');
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
@ -830,6 +830,100 @@ async function sendDeepSeekRequest(request, response) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sends a request to XAI API.
|
||||||
|
* @param {express.Request} request Express request
|
||||||
|
* @param {express.Response} response Express response
|
||||||
|
*/
|
||||||
|
async function sendXaiRequest(request, response) {
|
||||||
|
const apiUrl = new URL(request.body.reverse_proxy || API_XAI).toString();
|
||||||
|
const apiKey = request.body.reverse_proxy ? request.body.proxy_password : readSecret(request.user.directories, SECRET_KEYS.XAI);
|
||||||
|
|
||||||
|
if (!apiKey && !request.body.reverse_proxy) {
|
||||||
|
console.warn('xAI API key is missing.');
|
||||||
|
return response.status(400).send({ error: true });
|
||||||
|
}
|
||||||
|
|
||||||
|
const controller = new AbortController();
|
||||||
|
request.socket.removeAllListeners('close');
|
||||||
|
request.socket.on('close', function () {
|
||||||
|
controller.abort();
|
||||||
|
});
|
||||||
|
|
||||||
|
try {
|
||||||
|
let bodyParams = {};
|
||||||
|
|
||||||
|
if (request.body.logprobs > 0) {
|
||||||
|
bodyParams['top_logprobs'] = request.body.logprobs;
|
||||||
|
bodyParams['logprobs'] = true;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (Array.isArray(request.body.tools) && request.body.tools.length > 0) {
|
||||||
|
bodyParams['tools'] = request.body.tools;
|
||||||
|
bodyParams['tool_choice'] = request.body.tool_choice;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (Array.isArray(request.body.stop) && request.body.stop.length > 0) {
|
||||||
|
bodyParams['stop'] = request.body.stop;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (['grok-3-mini-beta', 'grok-3-mini-fast-beta'].includes(request.body.model)) {
|
||||||
|
bodyParams['reasoning_effort'] = request.body.reasoning_effort === 'high' ? 'high' : 'low';
|
||||||
|
}
|
||||||
|
|
||||||
|
const processedMessages = request.body.messages = convertXAIMessages(request.body.messages, getPromptNames(request));
|
||||||
|
|
||||||
|
const requestBody = {
|
||||||
|
'messages': processedMessages,
|
||||||
|
'model': request.body.model,
|
||||||
|
'temperature': request.body.temperature,
|
||||||
|
'max_tokens': request.body.max_tokens,
|
||||||
|
'max_completion_tokens': request.body.max_completion_tokens,
|
||||||
|
'stream': request.body.stream,
|
||||||
|
'presence_penalty': request.body.presence_penalty,
|
||||||
|
'frequency_penalty': request.body.frequency_penalty,
|
||||||
|
'top_p': request.body.top_p,
|
||||||
|
'seed': request.body.seed,
|
||||||
|
'n': request.body.n,
|
||||||
|
...bodyParams,
|
||||||
|
};
|
||||||
|
|
||||||
|
const config = {
|
||||||
|
method: 'POST',
|
||||||
|
headers: {
|
||||||
|
'Content-Type': 'application/json',
|
||||||
|
'Authorization': 'Bearer ' + apiKey,
|
||||||
|
},
|
||||||
|
body: JSON.stringify(requestBody),
|
||||||
|
signal: controller.signal,
|
||||||
|
};
|
||||||
|
|
||||||
|
console.debug('xAI request:', requestBody);
|
||||||
|
|
||||||
|
const generateResponse = await fetch(apiUrl + '/chat/completions', config);
|
||||||
|
|
||||||
|
if (request.body.stream) {
|
||||||
|
forwardFetchResponse(generateResponse, response);
|
||||||
|
} else {
|
||||||
|
if (!generateResponse.ok) {
|
||||||
|
const errorText = await generateResponse.text();
|
||||||
|
console.warn(`xAI API returned error: ${generateResponse.status} ${generateResponse.statusText} ${errorText}`);
|
||||||
|
const errorJson = tryParse(errorText) ?? { error: true };
|
||||||
|
return response.status(500).send(errorJson);
|
||||||
|
}
|
||||||
|
const generateResponseJson = await generateResponse.json();
|
||||||
|
console.debug('xAI response:', generateResponseJson);
|
||||||
|
return response.send(generateResponseJson);
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
console.error('Error communicating with xAI API: ', error);
|
||||||
|
if (!response.headersSent) {
|
||||||
|
response.send({ error: true });
|
||||||
|
} else {
|
||||||
|
response.end();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
export const router = express.Router();
|
export const router = express.Router();
|
||||||
|
|
||||||
@ -875,8 +969,9 @@ router.post('/status', async function (request, response_getstatus_openai) {
|
|||||||
api_key_openai = request.body.reverse_proxy ? request.body.proxy_password : readSecret(request.user.directories, SECRET_KEYS.DEEPSEEK);
|
api_key_openai = request.body.reverse_proxy ? request.body.proxy_password : readSecret(request.user.directories, SECRET_KEYS.DEEPSEEK);
|
||||||
headers = {};
|
headers = {};
|
||||||
} else if (request.body.chat_completion_source === CHAT_COMPLETION_SOURCES.XAI) {
|
} else if (request.body.chat_completion_source === CHAT_COMPLETION_SOURCES.XAI) {
|
||||||
api_url = API_XAI;
|
api_url = new URL(request.body.reverse_proxy || API_XAI);
|
||||||
api_key_openai = readSecret(request.user.directories, SECRET_KEYS.XAI);
|
api_key_openai = request.body.reverse_proxy ? request.body.proxy_password : readSecret(request.user.directories, SECRET_KEYS.XAI);
|
||||||
|
headers = {};
|
||||||
} else {
|
} else {
|
||||||
console.warn('This chat completion source is not supported yet.');
|
console.warn('This chat completion source is not supported yet.');
|
||||||
return response_getstatus_openai.status(400).send({ error: true });
|
return response_getstatus_openai.status(400).send({ error: true });
|
||||||
@ -1044,6 +1139,7 @@ router.post('/generate', function (request, response) {
|
|||||||
case CHAT_COMPLETION_SOURCES.MISTRALAI: return sendMistralAIRequest(request, response);
|
case CHAT_COMPLETION_SOURCES.MISTRALAI: return sendMistralAIRequest(request, response);
|
||||||
case CHAT_COMPLETION_SOURCES.COHERE: return sendCohereRequest(request, response);
|
case CHAT_COMPLETION_SOURCES.COHERE: return sendCohereRequest(request, response);
|
||||||
case CHAT_COMPLETION_SOURCES.DEEPSEEK: return sendDeepSeekRequest(request, response);
|
case CHAT_COMPLETION_SOURCES.DEEPSEEK: return sendDeepSeekRequest(request, response);
|
||||||
|
case CHAT_COMPLETION_SOURCES.XAI: return sendXaiRequest(request, response);
|
||||||
}
|
}
|
||||||
|
|
||||||
let apiUrl;
|
let apiUrl;
|
||||||
@ -1155,12 +1251,6 @@ router.post('/generate', function (request, response) {
|
|||||||
apiKey = readSecret(request.user.directories, SECRET_KEYS.ZEROONEAI);
|
apiKey = readSecret(request.user.directories, SECRET_KEYS.ZEROONEAI);
|
||||||
headers = {};
|
headers = {};
|
||||||
bodyParams = {};
|
bodyParams = {};
|
||||||
} else if (request.body.chat_completion_source === CHAT_COMPLETION_SOURCES.XAI) {
|
|
||||||
apiUrl = API_XAI;
|
|
||||||
apiKey = readSecret(request.user.directories, SECRET_KEYS.XAI);
|
|
||||||
headers = {};
|
|
||||||
bodyParams = {};
|
|
||||||
request.body.messages = convertXAIMessages(request.body.messages, getPromptNames(request));
|
|
||||||
} else {
|
} else {
|
||||||
console.warn('This chat completion source is not supported yet.');
|
console.warn('This chat completion source is not supported yet.');
|
||||||
return response.status(400).send({ error: true });
|
return response.status(400).send({ error: true });
|
||||||
@ -1173,12 +1263,6 @@ router.post('/generate', function (request, response) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if ([CHAT_COMPLETION_SOURCES.XAI].includes(request.body.chat_completion_source)) {
|
|
||||||
if (['grok-3-mini-beta', 'grok-3-mini-fast-beta'].includes(request.body.model)) {
|
|
||||||
bodyParams['reasoning_effort'] = request.body.reasoning_effort === 'high' ? 'high' : 'low';
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!apiKey && !request.body.reverse_proxy && request.body.chat_completion_source !== CHAT_COMPLETION_SOURCES.CUSTOM) {
|
if (!apiKey && !request.body.reverse_proxy && request.body.chat_completion_source !== CHAT_COMPLETION_SOURCES.CUSTOM) {
|
||||||
console.warn('OpenAI API key is missing.');
|
console.warn('OpenAI API key is missing.');
|
||||||
return response.status(400).send({ error: true });
|
return response.status(400).send({ error: true });
|
||||||
|
Reference in New Issue
Block a user