mirror of
https://github.com/SillyTavern/SillyTavern.git
synced 2025-06-05 21:59:27 +02:00
Replace use_[source] with chat_completion_source
Same as the is_[api] replacement--it's easier to have one enum field than several mutually-exclusive boolean ones
This commit is contained in:
29
server.js
29
server.js
@@ -2794,7 +2794,7 @@ app.post('/getstatus_openai', jsonParser, async function (request, response_gets
|
||||
let api_key_openai;
|
||||
let headers;
|
||||
|
||||
if (request.body.use_openrouter == false) {
|
||||
if (request.body.chat_completion_source !== 'openrouter') {
|
||||
api_url = new URL(request.body.reverse_proxy || API_OPENAI).toString();
|
||||
api_key_openai = request.body.reverse_proxy ? request.body.proxy_password : readSecret(SECRET_KEYS.OPENAI);
|
||||
headers = {};
|
||||
@@ -2822,7 +2822,7 @@ app.post('/getstatus_openai', jsonParser, async function (request, response_gets
|
||||
const data = await response.json();
|
||||
response_getstatus_openai.send(data);
|
||||
|
||||
if (request.body.use_openrouter && Array.isArray(data?.data)) {
|
||||
if (request.body.chat_completion_source === 'openrouter' && Array.isArray(data?.data)) {
|
||||
let models = [];
|
||||
|
||||
data.data.forEach(model => {
|
||||
@@ -3237,20 +3237,11 @@ async function sendPalmRequest(request, response) {
|
||||
app.post('/generate_openai', jsonParser, function (request, response_generate_openai) {
|
||||
if (!request.body) return response_generate_openai.status(400).send({ error: true });
|
||||
|
||||
if (request.body.use_claude) {
|
||||
return sendClaudeRequest(request, response_generate_openai);
|
||||
}
|
||||
|
||||
if (request.body.use_scale) {
|
||||
return sendScaleRequest(request, response_generate_openai);
|
||||
}
|
||||
|
||||
if (request.body.use_ai21) {
|
||||
return sendAI21Request(request, response_generate_openai);
|
||||
}
|
||||
|
||||
if (request.body.use_palm) {
|
||||
return sendPalmRequest(request, response_generate_openai);
|
||||
switch (request.body.chat_completion_source) {
|
||||
case 'claude': return sendClaudeRequest(request, response_generate_openai);
|
||||
case 'scale': return sendScaleRequest(request, response_generate_openai);
|
||||
case 'ai21': return sendAI21Request(request, response_generate_openai);
|
||||
case 'palm': return sendPalmRequest(request, response_generate_openai);
|
||||
}
|
||||
|
||||
let api_url;
|
||||
@@ -3258,7 +3249,7 @@ app.post('/generate_openai', jsonParser, function (request, response_generate_op
|
||||
let headers;
|
||||
let bodyParams;
|
||||
|
||||
if (!request.body.use_openrouter) {
|
||||
if (request.body.chat_completion_source !== 'openrouter') {
|
||||
api_url = new URL(request.body.reverse_proxy || API_OPENAI).toString();
|
||||
api_key_openai = request.body.reverse_proxy ? request.body.proxy_password : readSecret(SECRET_KEYS.OPENAI);
|
||||
headers = {};
|
||||
@@ -3290,7 +3281,9 @@ app.post('/generate_openai', jsonParser, function (request, response_generate_op
|
||||
|
||||
const isTextCompletion = Boolean(request.body.model && TEXT_COMPLETION_MODELS.includes(request.body.model)) || typeof request.body.messages === 'string';
|
||||
const textPrompt = isTextCompletion ? convertChatMLPrompt(request.body.messages) : '';
|
||||
const endpointUrl = isTextCompletion && !request.body.use_openrouter ? `${api_url}/completions` : `${api_url}/chat/completions`;
|
||||
const endpointUrl = isTextCompletion && request.body.chat_completion_source !== 'openrouter' ?
|
||||
`${api_url}/completions` :
|
||||
`${api_url}/chat/completions`;
|
||||
|
||||
const controller = new AbortController();
|
||||
request.socket.removeAllListeners('close');
|
||||
|
Reference in New Issue
Block a user