Replace use_[source] with chat_completion_source

Same as the is_[api] replacement--it's easier to have one enum field
than several mutually-exclusive boolean ones
This commit is contained in:
valadaptive
2023-12-03 08:31:58 -05:00
parent a68505e857
commit e33c8bd955
2 changed files with 14 additions and 25 deletions

View File

@@ -2794,7 +2794,7 @@ app.post('/getstatus_openai', jsonParser, async function (request, response_gets
let api_key_openai;
let headers;
if (request.body.use_openrouter == false) {
if (request.body.chat_completion_source !== 'openrouter') {
api_url = new URL(request.body.reverse_proxy || API_OPENAI).toString();
api_key_openai = request.body.reverse_proxy ? request.body.proxy_password : readSecret(SECRET_KEYS.OPENAI);
headers = {};
@@ -2822,7 +2822,7 @@ app.post('/getstatus_openai', jsonParser, async function (request, response_gets
const data = await response.json();
response_getstatus_openai.send(data);
if (request.body.use_openrouter && Array.isArray(data?.data)) {
if (request.body.chat_completion_source === 'openrouter' && Array.isArray(data?.data)) {
let models = [];
data.data.forEach(model => {
@@ -3237,20 +3237,11 @@ async function sendPalmRequest(request, response) {
app.post('/generate_openai', jsonParser, function (request, response_generate_openai) {
if (!request.body) return response_generate_openai.status(400).send({ error: true });
if (request.body.use_claude) {
return sendClaudeRequest(request, response_generate_openai);
}
if (request.body.use_scale) {
return sendScaleRequest(request, response_generate_openai);
}
if (request.body.use_ai21) {
return sendAI21Request(request, response_generate_openai);
}
if (request.body.use_palm) {
return sendPalmRequest(request, response_generate_openai);
switch (request.body.chat_completion_source) {
case 'claude': return sendClaudeRequest(request, response_generate_openai);
case 'scale': return sendScaleRequest(request, response_generate_openai);
case 'ai21': return sendAI21Request(request, response_generate_openai);
case 'palm': return sendPalmRequest(request, response_generate_openai);
}
let api_url;
@@ -3258,7 +3249,7 @@ app.post('/generate_openai', jsonParser, function (request, response_generate_op
let headers;
let bodyParams;
if (!request.body.use_openrouter) {
if (request.body.chat_completion_source !== 'openrouter') {
api_url = new URL(request.body.reverse_proxy || API_OPENAI).toString();
api_key_openai = request.body.reverse_proxy ? request.body.proxy_password : readSecret(SECRET_KEYS.OPENAI);
headers = {};
@@ -3290,7 +3281,9 @@ app.post('/generate_openai', jsonParser, function (request, response_generate_op
const isTextCompletion = Boolean(request.body.model && TEXT_COMPLETION_MODELS.includes(request.body.model)) || typeof request.body.messages === 'string';
const textPrompt = isTextCompletion ? convertChatMLPrompt(request.body.messages) : '';
const endpointUrl = isTextCompletion && !request.body.use_openrouter ? `${api_url}/completions` : `${api_url}/chat/completions`;
const endpointUrl = isTextCompletion && request.body.chat_completion_source !== 'openrouter' ?
`${api_url}/completions` :
`${api_url}/chat/completions`;
const controller = new AbortController();
request.socket.removeAllListeners('close');