mirror of
https://github.com/SillyTavern/SillyTavern.git
synced 2025-04-17 12:17:21 +02:00
Merge branch 'staging' of https://github.com/Cohee1207/SillyTavern into staging
This commit is contained in:
commit
b0907d23cc
4
package-lock.json
generated
4
package-lock.json
generated
@ -1,12 +1,12 @@
|
|||||||
{
|
{
|
||||||
"name": "sillytavern",
|
"name": "sillytavern",
|
||||||
"version": "1.11.0",
|
"version": "1.11.1",
|
||||||
"lockfileVersion": 3,
|
"lockfileVersion": 3,
|
||||||
"requires": true,
|
"requires": true,
|
||||||
"packages": {
|
"packages": {
|
||||||
"": {
|
"": {
|
||||||
"name": "sillytavern",
|
"name": "sillytavern",
|
||||||
"version": "1.11.0",
|
"version": "1.11.1",
|
||||||
"hasInstallScript": true,
|
"hasInstallScript": true,
|
||||||
"license": "AGPL-3.0",
|
"license": "AGPL-3.0",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
|
@ -51,7 +51,7 @@
|
|||||||
"type": "git",
|
"type": "git",
|
||||||
"url": "https://github.com/SillyTavern/SillyTavern.git"
|
"url": "https://github.com/SillyTavern/SillyTavern.git"
|
||||||
},
|
},
|
||||||
"version": "1.11.0",
|
"version": "1.11.1",
|
||||||
"scripts": {
|
"scripts": {
|
||||||
"start": "node server.js",
|
"start": "node server.js",
|
||||||
"start-multi": "node server.js --disableCsrf",
|
"start-multi": "node server.js --disableCsrf",
|
||||||
|
15
public/img/mistralai.svg
Normal file
15
public/img/mistralai.svg
Normal file
@ -0,0 +1,15 @@
|
|||||||
|
<?xml version="1.0" standalone="no"?>
|
||||||
|
<!DOCTYPE svg PUBLIC "-//W3C//DTD SVG 20010904//EN"
|
||||||
|
"http://www.w3.org/TR/2001/REC-SVG-20010904/DTD/svg10.dtd">
|
||||||
|
<svg version="1.0" xmlns="http://www.w3.org/2000/svg"
|
||||||
|
width="64.000000pt" height="64.000000pt" viewBox="0 0 53.000000 60.000000"
|
||||||
|
preserveAspectRatio="xMidYMid meet">
|
||||||
|
|
||||||
|
<g transform="translate(0.000000,63.000000) scale(0.100000,-0.100000)"
|
||||||
|
fill="#000000" stroke="none">
|
||||||
|
<path d="M40 320 l0 -240 70 0 70 0 0 95 c0 95 0 95 25 95 23 0 25 -3 25 -50
|
||||||
|
l0 -50 70 0 70 0 0 50 c0 47 2 50 25 50 25 0 25 0 25 -95 l0 -95 70 0 70 0 0
|
||||||
|
240 0 240 -70 0 -70 0 0 -44 0 -45 -47 -3 -48 -3 -3 -47 c-3 -43 -5 -48 -28
|
||||||
|
-48 -22 0 -24 4 -24 50 l0 50 -45 0 -45 0 0 45 0 45 -70 0 -70 0 0 -240z"/>
|
||||||
|
</g>
|
||||||
|
</svg>
|
After Width: | Height: | Size: 731 B |
@ -444,7 +444,7 @@
|
|||||||
complete.</span>
|
complete.</span>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
<div class="range-block" data-source="openai,claude,windowai,openrouter,ai21,scale,makersuite">
|
<div class="range-block" data-source="openai,claude,windowai,openrouter,ai21,scale,makersuite,mistralai">
|
||||||
<div class="range-block-title" data-i18n="Temperature">
|
<div class="range-block-title" data-i18n="Temperature">
|
||||||
Temperature
|
Temperature
|
||||||
</div>
|
</div>
|
||||||
@ -509,7 +509,7 @@
|
|||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
<div data-newbie-hidden class="range-block" data-source="openai,claude,openrouter,ai21,scale,makersuite">
|
<div data-newbie-hidden class="range-block" data-source="openai,claude,openrouter,ai21,scale,makersuite,mistralai">
|
||||||
<div class="range-block-title" data-i18n="Top-p">
|
<div class="range-block-title" data-i18n="Top-p">
|
||||||
Top P
|
Top P
|
||||||
</div>
|
</div>
|
||||||
@ -749,7 +749,7 @@
|
|||||||
<div id="openai_proxy_password_show" title="Peek a password" class="menu_button fa-solid fa-eye-slash fa-fw"></div>
|
<div id="openai_proxy_password_show" title="Peek a password" class="menu_button fa-solid fa-eye-slash fa-fw"></div>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
<div data-newbie-hidden class="range-block" data-source="openai,openrouter">
|
<div data-newbie-hidden class="range-block" data-source="openai,openrouter,mistralai">
|
||||||
<div class="range-block-title justifyLeft" data-i18n="Seed">
|
<div class="range-block-title justifyLeft" data-i18n="Seed">
|
||||||
Seed
|
Seed
|
||||||
</div>
|
</div>
|
||||||
@ -1593,7 +1593,7 @@
|
|||||||
<option value="koboldhorde"><span data-i18n="KoboldAI Horde">KoboldAI Horde</span></option>
|
<option value="koboldhorde"><span data-i18n="KoboldAI Horde">KoboldAI Horde</span></option>
|
||||||
<option value="novel"><span data-i18n="NovelAI">NovelAI</span></option>
|
<option value="novel"><span data-i18n="NovelAI">NovelAI</span></option>
|
||||||
<option value="textgenerationwebui"><span data-i18n="Text Completion">Text Completion (ooba, Mancer, Aphrodite, TabbyAPI, KoboldCpp)</span></option>
|
<option value="textgenerationwebui"><span data-i18n="Text Completion">Text Completion (ooba, Mancer, Aphrodite, TabbyAPI, KoboldCpp)</span></option>
|
||||||
<option value="openai"><span data-i18n="Chat Completion (OpenAI, Claude, Window/OpenRouter, Scale, AI21)">Chat Completion (OpenAI, Claude, Window, OpenRouter, Scale, AI21, Google MakerSuite)</span></option>
|
<option value="openai"><span data-i18n="Chat Completion (OpenAI, Claude, Window/OpenRouter, Scale, AI21, Google MakerSuite, MistralAI)">Chat Completion (OpenAI, Claude, Window, OpenRouter, Scale, AI21, Google MakerSuite, MistralAI)</span></option>
|
||||||
</select>
|
</select>
|
||||||
</div>
|
</div>
|
||||||
<div id="kobold_horde" style="position: relative;"> <!-- shows the kobold settings -->
|
<div id="kobold_horde" style="position: relative;"> <!-- shows the kobold settings -->
|
||||||
@ -1842,6 +1842,7 @@
|
|||||||
<option value="scale">Scale</option>
|
<option value="scale">Scale</option>
|
||||||
<option value="ai21">AI21</option>
|
<option value="ai21">AI21</option>
|
||||||
<option value="makersuite">Google MakerSuite</option>
|
<option value="makersuite">Google MakerSuite</option>
|
||||||
|
<option value="mistralai">MistralAI</option>
|
||||||
</select>
|
</select>
|
||||||
<form id="openai_form" data-source="openai" action="javascript:void(null);" method="post" enctype="multipart/form-data">
|
<form id="openai_form" data-source="openai" action="javascript:void(null);" method="post" enctype="multipart/form-data">
|
||||||
<h4><span data-i18n="OpenAI API key">OpenAI API key</span></h4>
|
<h4><span data-i18n="OpenAI API key">OpenAI API key</span></h4>
|
||||||
@ -2126,6 +2127,26 @@
|
|||||||
</select>
|
</select>
|
||||||
</div>
|
</div>
|
||||||
</form>
|
</form>
|
||||||
|
<form id="mistralai_form" data-source="mistralai" action="javascript:void(null);" method="post" enctype="multipart/form-data">
|
||||||
|
<h4 data-i18n="MistralAI API Key">MistralAI API Key</h4>
|
||||||
|
<div class="flex-container">
|
||||||
|
<input id="api_key_mistralai" name="api_key_mistralai" class="text_pole flex1" maxlength="500" value="" type="text" autocomplete="off">
|
||||||
|
<div title="Clear your API key" data-i18n="[title]Clear your API key" class="menu_button fa-solid fa-circle-xmark clear-api-key" data-key="api_key_mistralai"></div>
|
||||||
|
</div>
|
||||||
|
<div data-for="api_key_mistralai" class="neutral_warning">
|
||||||
|
For privacy reasons, your API key will be hidden after you reload the page.
|
||||||
|
</div>
|
||||||
|
<div>
|
||||||
|
<h4 data-i18n="MistralAI Model">MistralAI Model</h4>
|
||||||
|
<select id="model_mistralai_select">
|
||||||
|
<optgroup label="Latest">
|
||||||
|
<option value="mistral-tiny">mistral-tiny</option>
|
||||||
|
<option value="mistral-small">mistral-small</option>
|
||||||
|
<option value="mistral-medium">mistral-medium</option>
|
||||||
|
</optgroup>
|
||||||
|
</select>
|
||||||
|
</div>
|
||||||
|
</form>
|
||||||
<div class="flex-container flex">
|
<div class="flex-container flex">
|
||||||
<div id="api_button_openai" class="api_button menu_button menu_button_icon" type="submit" data-i18n="Connect">Connect</div>
|
<div id="api_button_openai" class="api_button menu_button menu_button_icon" type="submit" data-i18n="Connect">Connect</div>
|
||||||
<div class="api_loading menu_button" data-i18n="Cancel">Cancel</div>
|
<div class="api_loading menu_button" data-i18n="Cancel">Cancel</div>
|
||||||
|
@ -5396,6 +5396,7 @@ function changeMainAPI() {
|
|||||||
case chat_completion_sources.OPENAI:
|
case chat_completion_sources.OPENAI:
|
||||||
case chat_completion_sources.AI21:
|
case chat_completion_sources.AI21:
|
||||||
case chat_completion_sources.MAKERSUITE:
|
case chat_completion_sources.MAKERSUITE:
|
||||||
|
case chat_completion_sources.MISTRALAI:
|
||||||
default:
|
default:
|
||||||
setupChatCompletionPromptManager(oai_settings);
|
setupChatCompletionPromptManager(oai_settings);
|
||||||
break;
|
break;
|
||||||
@ -7546,6 +7547,11 @@ async function connectAPISlash(_, text) {
|
|||||||
source: 'makersuite',
|
source: 'makersuite',
|
||||||
button: '#api_button_openai',
|
button: '#api_button_openai',
|
||||||
},
|
},
|
||||||
|
'mistralai': {
|
||||||
|
selected: 'openai',
|
||||||
|
source: 'mistralai',
|
||||||
|
button: '#api_button_openai',
|
||||||
|
},
|
||||||
};
|
};
|
||||||
|
|
||||||
const apiConfig = apiMap[text.toLowerCase()];
|
const apiConfig = apiMap[text.toLowerCase()];
|
||||||
@ -7832,7 +7838,7 @@ jQuery(async function () {
|
|||||||
}
|
}
|
||||||
|
|
||||||
registerSlashCommand('dupe', DupeChar, [], '– duplicates the currently selected character', true, true);
|
registerSlashCommand('dupe', DupeChar, [], '– duplicates the currently selected character', true, true);
|
||||||
registerSlashCommand('api', connectAPISlash, [], '<span class="monospace">(kobold, horde, novel, ooba, tabby, mancer, aphrodite, kcpp, oai, claude, windowai, openrouter, scale, ai21, makersuite)</span> – connect to an API', true, true);
|
registerSlashCommand('api', connectAPISlash, [], '<span class="monospace">(kobold, horde, novel, ooba, tabby, mancer, aphrodite, kcpp, oai, claude, windowai, openrouter, scale, ai21, makersuite, mistralai)</span> – connect to an API', true, true);
|
||||||
registerSlashCommand('impersonate', doImpersonate, ['imp'], '– calls an impersonation response', true, true);
|
registerSlashCommand('impersonate', doImpersonate, ['imp'], '– calls an impersonation response', true, true);
|
||||||
registerSlashCommand('delchat', doDeleteChat, [], '– deletes the current chat', true, true);
|
registerSlashCommand('delchat', doDeleteChat, [], '– deletes the current chat', true, true);
|
||||||
registerSlashCommand('closechat', doCloseChat, [], '– closes the current chat', true, true);
|
registerSlashCommand('closechat', doCloseChat, [], '– closes the current chat', true, true);
|
||||||
|
@ -396,6 +396,7 @@ function RA_autoconnect(PrevApi) {
|
|||||||
|| (secret_state[SECRET_KEYS.OPENROUTER] && oai_settings.chat_completion_source == chat_completion_sources.OPENROUTER)
|
|| (secret_state[SECRET_KEYS.OPENROUTER] && oai_settings.chat_completion_source == chat_completion_sources.OPENROUTER)
|
||||||
|| (secret_state[SECRET_KEYS.AI21] && oai_settings.chat_completion_source == chat_completion_sources.AI21)
|
|| (secret_state[SECRET_KEYS.AI21] && oai_settings.chat_completion_source == chat_completion_sources.AI21)
|
||||||
|| (secret_state[SECRET_KEYS.MAKERSUITE] && oai_settings.chat_completion_source == chat_completion_sources.MAKERSUITE)
|
|| (secret_state[SECRET_KEYS.MAKERSUITE] && oai_settings.chat_completion_source == chat_completion_sources.MAKERSUITE)
|
||||||
|
|| (secret_state[SECRET_KEYS.MISTRALAI] && oai_settings.chat_completion_source == chat_completion_sources.MISTRALAI)
|
||||||
) {
|
) {
|
||||||
$('#api_button_openai').trigger('click');
|
$('#api_button_openai').trigger('click');
|
||||||
}
|
}
|
||||||
|
@ -1711,7 +1711,7 @@ async function getPrompt(generationType, message, trigger, quietPrompt) {
|
|||||||
prompt = message || getRawLastMessage();
|
prompt = message || getRawLastMessage();
|
||||||
break;
|
break;
|
||||||
case generationMode.FREE:
|
case generationMode.FREE:
|
||||||
prompt = trigger.trim();
|
prompt = generateFreeModePrompt(trigger.trim());
|
||||||
break;
|
break;
|
||||||
case generationMode.FACE_MULTIMODAL:
|
case generationMode.FACE_MULTIMODAL:
|
||||||
case generationMode.CHARACTER_MULTIMODAL:
|
case generationMode.CHARACTER_MULTIMODAL:
|
||||||
@ -1730,6 +1730,36 @@ async function getPrompt(generationType, message, trigger, quietPrompt) {
|
|||||||
return prompt;
|
return prompt;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Generates a free prompt with a character-specific prompt prefix support.
|
||||||
|
* @param {string} trigger - The prompt to use for the image generation.
|
||||||
|
* @returns {string}
|
||||||
|
*/
|
||||||
|
function generateFreeModePrompt(trigger) {
|
||||||
|
return trigger
|
||||||
|
.replace(/(?:^char(\s|,)|\{\{charPrefix\}\})/gi, (_, suffix) => {
|
||||||
|
const getLastCharacterKey = () => {
|
||||||
|
if (typeof this_chid !== 'undefined') {
|
||||||
|
return getCharaFilename(this_chid);
|
||||||
|
}
|
||||||
|
const context = getContext();
|
||||||
|
for (let i = context.chat.length - 1; i >= 0; i--) {
|
||||||
|
const message = context.chat[i];
|
||||||
|
if (message.is_user || message.is_system) {
|
||||||
|
continue;
|
||||||
|
} else if (typeof message.original_avatar === 'string') {
|
||||||
|
return message.original_avatar.replace(/\.[^/.]+$/, '');
|
||||||
|
}
|
||||||
|
}
|
||||||
|
throw new Error('No usable messages found.');
|
||||||
|
};
|
||||||
|
|
||||||
|
const key = getLastCharacterKey();
|
||||||
|
const value = (extension_settings.sd.character_prompts[key] || '').trim();
|
||||||
|
return value ? value + (suffix || '') : '';
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Generates a prompt using multimodal captioning.
|
* Generates a prompt using multimodal captioning.
|
||||||
* @param {number} generationType - The type of image generation to perform.
|
* @param {number} generationType - The type of image generation to perform.
|
||||||
|
@ -394,7 +394,8 @@ async function getSavedHashes(collectionId) {
|
|||||||
*/
|
*/
|
||||||
async function insertVectorItems(collectionId, items) {
|
async function insertVectorItems(collectionId, items) {
|
||||||
if (settings.source === 'openai' && !secret_state[SECRET_KEYS.OPENAI] ||
|
if (settings.source === 'openai' && !secret_state[SECRET_KEYS.OPENAI] ||
|
||||||
settings.source === 'palm' && !secret_state[SECRET_KEYS.MAKERSUITE]) {
|
settings.source === 'palm' && !secret_state[SECRET_KEYS.MAKERSUITE] ||
|
||||||
|
settings.source === 'mistral' && !secret_state[SECRET_KEYS.MISTRALAI]) {
|
||||||
throw new Error('Vectors: API key missing', { cause: 'api_key_missing' });
|
throw new Error('Vectors: API key missing', { cause: 'api_key_missing' });
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -13,6 +13,7 @@
|
|||||||
<option value="transformers">Local (Transformers)</option>
|
<option value="transformers">Local (Transformers)</option>
|
||||||
<option value="openai">OpenAI</option>
|
<option value="openai">OpenAI</option>
|
||||||
<option value="palm">Google MakerSuite (PaLM)</option>
|
<option value="palm">Google MakerSuite (PaLM)</option>
|
||||||
|
<option value="mistral">MistralAI</option>
|
||||||
</select>
|
</select>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
|
@ -164,6 +164,7 @@ export const chat_completion_sources = {
|
|||||||
OPENROUTER: 'openrouter',
|
OPENROUTER: 'openrouter',
|
||||||
AI21: 'ai21',
|
AI21: 'ai21',
|
||||||
MAKERSUITE: 'makersuite',
|
MAKERSUITE: 'makersuite',
|
||||||
|
MISTRALAI: 'mistralai',
|
||||||
};
|
};
|
||||||
|
|
||||||
const prefixMap = selected_group ? {
|
const prefixMap = selected_group ? {
|
||||||
@ -208,6 +209,7 @@ const default_settings = {
|
|||||||
claude_model: 'claude-instant-v1',
|
claude_model: 'claude-instant-v1',
|
||||||
google_model: 'gemini-pro',
|
google_model: 'gemini-pro',
|
||||||
ai21_model: 'j2-ultra',
|
ai21_model: 'j2-ultra',
|
||||||
|
mistralai_model: 'mistral-medium',
|
||||||
windowai_model: '',
|
windowai_model: '',
|
||||||
openrouter_model: openrouter_website_model,
|
openrouter_model: openrouter_website_model,
|
||||||
openrouter_use_fallback: false,
|
openrouter_use_fallback: false,
|
||||||
@ -263,6 +265,7 @@ const oai_settings = {
|
|||||||
claude_model: 'claude-instant-v1',
|
claude_model: 'claude-instant-v1',
|
||||||
google_model: 'gemini-pro',
|
google_model: 'gemini-pro',
|
||||||
ai21_model: 'j2-ultra',
|
ai21_model: 'j2-ultra',
|
||||||
|
mistralai_model: 'mistral-medium',
|
||||||
windowai_model: '',
|
windowai_model: '',
|
||||||
openrouter_model: openrouter_website_model,
|
openrouter_model: openrouter_website_model,
|
||||||
openrouter_use_fallback: false,
|
openrouter_use_fallback: false,
|
||||||
@ -1261,6 +1264,8 @@ function getChatCompletionModel() {
|
|||||||
return oai_settings.openrouter_model !== openrouter_website_model ? oai_settings.openrouter_model : null;
|
return oai_settings.openrouter_model !== openrouter_website_model ? oai_settings.openrouter_model : null;
|
||||||
case chat_completion_sources.AI21:
|
case chat_completion_sources.AI21:
|
||||||
return oai_settings.ai21_model;
|
return oai_settings.ai21_model;
|
||||||
|
case chat_completion_sources.MISTRALAI:
|
||||||
|
return oai_settings.mistralai_model;
|
||||||
default:
|
default:
|
||||||
throw new Error(`Unknown chat completion source: ${oai_settings.chat_completion_source}`);
|
throw new Error(`Unknown chat completion source: ${oai_settings.chat_completion_source}`);
|
||||||
}
|
}
|
||||||
@ -1448,6 +1453,7 @@ async function sendOpenAIRequest(type, messages, signal) {
|
|||||||
const isAI21 = oai_settings.chat_completion_source == chat_completion_sources.AI21;
|
const isAI21 = oai_settings.chat_completion_source == chat_completion_sources.AI21;
|
||||||
const isGoogle = oai_settings.chat_completion_source == chat_completion_sources.MAKERSUITE;
|
const isGoogle = oai_settings.chat_completion_source == chat_completion_sources.MAKERSUITE;
|
||||||
const isOAI = oai_settings.chat_completion_source == chat_completion_sources.OPENAI;
|
const isOAI = oai_settings.chat_completion_source == chat_completion_sources.OPENAI;
|
||||||
|
const isMistral = oai_settings.chat_completion_source == chat_completion_sources.MISTRALAI;
|
||||||
const isTextCompletion = (isOAI && textCompletionModels.includes(oai_settings.openai_model)) || (isOpenRouter && oai_settings.openrouter_force_instruct && power_user.instruct.enabled);
|
const isTextCompletion = (isOAI && textCompletionModels.includes(oai_settings.openai_model)) || (isOpenRouter && oai_settings.openrouter_force_instruct && power_user.instruct.enabled);
|
||||||
const isQuiet = type === 'quiet';
|
const isQuiet = type === 'quiet';
|
||||||
const isImpersonate = type === 'impersonate';
|
const isImpersonate = type === 'impersonate';
|
||||||
@ -1555,7 +1561,11 @@ async function sendOpenAIRequest(type, messages, signal) {
|
|||||||
generate_data['stop_tokens'] = [name1 + ':', oai_settings.new_chat_prompt, oai_settings.new_group_chat_prompt];
|
generate_data['stop_tokens'] = [name1 + ':', oai_settings.new_chat_prompt, oai_settings.new_group_chat_prompt];
|
||||||
}
|
}
|
||||||
|
|
||||||
if ((isOAI || isOpenRouter) && oai_settings.seed >= 0) {
|
if (isMistral) {
|
||||||
|
generate_data['safe_mode'] = false; // already defaults to false, but just incase they change that in the future.
|
||||||
|
}
|
||||||
|
|
||||||
|
if ((isOAI || isOpenRouter || isMistral) && oai_settings.seed >= 0) {
|
||||||
generate_data['seed'] = oai_settings.seed;
|
generate_data['seed'] = oai_settings.seed;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -2300,6 +2310,7 @@ function loadOpenAISettings(data, settings) {
|
|||||||
oai_settings.openrouter_use_fallback = settings.openrouter_use_fallback ?? default_settings.openrouter_use_fallback;
|
oai_settings.openrouter_use_fallback = settings.openrouter_use_fallback ?? default_settings.openrouter_use_fallback;
|
||||||
oai_settings.openrouter_force_instruct = settings.openrouter_force_instruct ?? default_settings.openrouter_force_instruct;
|
oai_settings.openrouter_force_instruct = settings.openrouter_force_instruct ?? default_settings.openrouter_force_instruct;
|
||||||
oai_settings.ai21_model = settings.ai21_model ?? default_settings.ai21_model;
|
oai_settings.ai21_model = settings.ai21_model ?? default_settings.ai21_model;
|
||||||
|
oai_settings.mistralai_model = settings.mistralai_model ?? default_settings.mistralai_model;
|
||||||
oai_settings.google_model = settings.google_model ?? default_settings.google_model;
|
oai_settings.google_model = settings.google_model ?? default_settings.google_model;
|
||||||
oai_settings.chat_completion_source = settings.chat_completion_source ?? default_settings.chat_completion_source;
|
oai_settings.chat_completion_source = settings.chat_completion_source ?? default_settings.chat_completion_source;
|
||||||
oai_settings.api_url_scale = settings.api_url_scale ?? default_settings.api_url_scale;
|
oai_settings.api_url_scale = settings.api_url_scale ?? default_settings.api_url_scale;
|
||||||
@ -2342,6 +2353,8 @@ function loadOpenAISettings(data, settings) {
|
|||||||
$(`#model_google_select option[value="${oai_settings.google_model}"`).attr('selected', true);
|
$(`#model_google_select option[value="${oai_settings.google_model}"`).attr('selected', true);
|
||||||
$('#model_ai21_select').val(oai_settings.ai21_model);
|
$('#model_ai21_select').val(oai_settings.ai21_model);
|
||||||
$(`#model_ai21_select option[value="${oai_settings.ai21_model}"`).attr('selected', true);
|
$(`#model_ai21_select option[value="${oai_settings.ai21_model}"`).attr('selected', true);
|
||||||
|
$('#model_mistralai_select').val(oai_settings.mistralai_model);
|
||||||
|
$(`#model_mistralai_select option[value="${oai_settings.mistralai_model}"`).attr('selected', true);
|
||||||
$('#openai_max_context').val(oai_settings.openai_max_context);
|
$('#openai_max_context').val(oai_settings.openai_max_context);
|
||||||
$('#openai_max_context_counter').val(`${oai_settings.openai_max_context}`);
|
$('#openai_max_context_counter').val(`${oai_settings.openai_max_context}`);
|
||||||
$('#model_openrouter_select').val(oai_settings.openrouter_model);
|
$('#model_openrouter_select').val(oai_settings.openrouter_model);
|
||||||
@ -2449,7 +2462,7 @@ async function getStatusOpen() {
|
|||||||
chat_completion_source: oai_settings.chat_completion_source,
|
chat_completion_source: oai_settings.chat_completion_source,
|
||||||
};
|
};
|
||||||
|
|
||||||
if (oai_settings.reverse_proxy && oai_settings.chat_completion_source !== chat_completion_sources.OPENROUTER) {
|
if (oai_settings.reverse_proxy && (oai_settings.chat_completion_source === chat_completion_sources.OPENAI || oai_settings.chat_completion_source === chat_completion_sources.CLAUDE)) {
|
||||||
validateReverseProxy();
|
validateReverseProxy();
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -2519,6 +2532,7 @@ async function saveOpenAIPreset(name, settings, triggerUi = true) {
|
|||||||
openrouter_group_models: settings.openrouter_group_models,
|
openrouter_group_models: settings.openrouter_group_models,
|
||||||
openrouter_sort_models: settings.openrouter_sort_models,
|
openrouter_sort_models: settings.openrouter_sort_models,
|
||||||
ai21_model: settings.ai21_model,
|
ai21_model: settings.ai21_model,
|
||||||
|
mistralai_model: settings.mistralai_model,
|
||||||
google_model: settings.google_model,
|
google_model: settings.google_model,
|
||||||
temperature: settings.temp_openai,
|
temperature: settings.temp_openai,
|
||||||
frequency_penalty: settings.freq_pen_openai,
|
frequency_penalty: settings.freq_pen_openai,
|
||||||
@ -2890,6 +2904,7 @@ function onSettingsPresetChange() {
|
|||||||
openrouter_group_models: ['#openrouter_group_models', 'openrouter_group_models', false],
|
openrouter_group_models: ['#openrouter_group_models', 'openrouter_group_models', false],
|
||||||
openrouter_sort_models: ['#openrouter_sort_models', 'openrouter_sort_models', false],
|
openrouter_sort_models: ['#openrouter_sort_models', 'openrouter_sort_models', false],
|
||||||
ai21_model: ['#model_ai21_select', 'ai21_model', false],
|
ai21_model: ['#model_ai21_select', 'ai21_model', false],
|
||||||
|
mistralai_model: ['#model_mistralai_select', 'mistralai_model', false],
|
||||||
google_model: ['#model_google_select', 'google_model', false],
|
google_model: ['#model_google_select', 'google_model', false],
|
||||||
openai_max_context: ['#openai_max_context', 'openai_max_context', false],
|
openai_max_context: ['#openai_max_context', 'openai_max_context', false],
|
||||||
openai_max_tokens: ['#openai_max_tokens', 'openai_max_tokens', false],
|
openai_max_tokens: ['#openai_max_tokens', 'openai_max_tokens', false],
|
||||||
@ -3074,6 +3089,11 @@ async function onModelChange() {
|
|||||||
oai_settings.google_model = value;
|
oai_settings.google_model = value;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if ($(this).is('#model_mistralai_select')) {
|
||||||
|
console.log('MistralAI model changed to', value);
|
||||||
|
oai_settings.mistralai_model = value;
|
||||||
|
}
|
||||||
|
|
||||||
if (oai_settings.chat_completion_source == chat_completion_sources.SCALE) {
|
if (oai_settings.chat_completion_source == chat_completion_sources.SCALE) {
|
||||||
if (oai_settings.max_context_unlocked) {
|
if (oai_settings.max_context_unlocked) {
|
||||||
$('#openai_max_context').attr('max', unlocked_max);
|
$('#openai_max_context').attr('max', unlocked_max);
|
||||||
@ -3179,6 +3199,16 @@ async function onModelChange() {
|
|||||||
$('#temp_openai').attr('max', oai_max_temp).val(oai_settings.temp_openai).trigger('input');
|
$('#temp_openai').attr('max', oai_max_temp).val(oai_settings.temp_openai).trigger('input');
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (oai_settings.chat_completion_source === chat_completion_sources.MISTRALAI) {
|
||||||
|
$('#openai_max_context').attr('max', max_32k);
|
||||||
|
oai_settings.openai_max_context = Math.min(oai_settings.openai_max_context, Number($('#openai_max_context').attr('max')));
|
||||||
|
$('#openai_max_context').val(oai_settings.openai_max_context).trigger('input');
|
||||||
|
|
||||||
|
//mistral also caps temp at 1.0
|
||||||
|
oai_settings.temp_openai = Math.min(claude_max_temp, oai_settings.temp_openai);
|
||||||
|
$('#temp_openai').attr('max', claude_max_temp).val(oai_settings.temp_openai).trigger('input');
|
||||||
|
}
|
||||||
|
|
||||||
if (oai_settings.chat_completion_source == chat_completion_sources.AI21) {
|
if (oai_settings.chat_completion_source == chat_completion_sources.AI21) {
|
||||||
if (oai_settings.max_context_unlocked) {
|
if (oai_settings.max_context_unlocked) {
|
||||||
$('#openai_max_context').attr('max', unlocked_max);
|
$('#openai_max_context').attr('max', unlocked_max);
|
||||||
@ -3340,6 +3370,19 @@ async function onConnectButtonClick(e) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (oai_settings.chat_completion_source == chat_completion_sources.MISTRALAI) {
|
||||||
|
const api_key_mistralai = String($('#api_key_mistralai').val()).trim();
|
||||||
|
|
||||||
|
if (api_key_mistralai.length) {
|
||||||
|
await writeSecret(SECRET_KEYS.MISTRALAI, api_key_mistralai);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!secret_state[SECRET_KEYS.MISTRALAI]) {
|
||||||
|
console.log('No secret key saved for MistralAI');
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
startStatusLoading();
|
startStatusLoading();
|
||||||
saveSettingsDebounced();
|
saveSettingsDebounced();
|
||||||
await getStatusOpen();
|
await getStatusOpen();
|
||||||
@ -3372,6 +3415,9 @@ function toggleChatCompletionForms() {
|
|||||||
else if (oai_settings.chat_completion_source == chat_completion_sources.AI21) {
|
else if (oai_settings.chat_completion_source == chat_completion_sources.AI21) {
|
||||||
$('#model_ai21_select').trigger('change');
|
$('#model_ai21_select').trigger('change');
|
||||||
}
|
}
|
||||||
|
else if (oai_settings.chat_completion_source == chat_completion_sources.MISTRALAI) {
|
||||||
|
$('#model_mistralai_select').trigger('change');
|
||||||
|
}
|
||||||
$('[data-source]').each(function () {
|
$('[data-source]').each(function () {
|
||||||
const validSources = $(this).data('source').split(',');
|
const validSources = $(this).data('source').split(',');
|
||||||
$(this).toggle(validSources.includes(oai_settings.chat_completion_source));
|
$(this).toggle(validSources.includes(oai_settings.chat_completion_source));
|
||||||
@ -3749,6 +3795,7 @@ $(document).ready(async function () {
|
|||||||
$('#openrouter_group_models').on('change', onOpenrouterModelSortChange);
|
$('#openrouter_group_models').on('change', onOpenrouterModelSortChange);
|
||||||
$('#openrouter_sort_models').on('change', onOpenrouterModelSortChange);
|
$('#openrouter_sort_models').on('change', onOpenrouterModelSortChange);
|
||||||
$('#model_ai21_select').on('change', onModelChange);
|
$('#model_ai21_select').on('change', onModelChange);
|
||||||
|
$('#model_mistralai_select').on('change', onModelChange);
|
||||||
$('#settings_preset_openai').on('change', onSettingsPresetChange);
|
$('#settings_preset_openai').on('change', onSettingsPresetChange);
|
||||||
$('#new_oai_preset').on('click', onNewPresetClick);
|
$('#new_oai_preset').on('click', onNewPresetClick);
|
||||||
$('#delete_oai_preset').on('click', onDeletePresetClick);
|
$('#delete_oai_preset').on('click', onDeletePresetClick);
|
||||||
|
@ -14,6 +14,7 @@ export const SECRET_KEYS = {
|
|||||||
SCALE_COOKIE: 'scale_cookie',
|
SCALE_COOKIE: 'scale_cookie',
|
||||||
MAKERSUITE: 'api_key_makersuite',
|
MAKERSUITE: 'api_key_makersuite',
|
||||||
SERPAPI: 'api_key_serpapi',
|
SERPAPI: 'api_key_serpapi',
|
||||||
|
MISTRALAI: 'api_key_mistralai',
|
||||||
};
|
};
|
||||||
|
|
||||||
const INPUT_MAP = {
|
const INPUT_MAP = {
|
||||||
@ -29,6 +30,7 @@ const INPUT_MAP = {
|
|||||||
[SECRET_KEYS.MAKERSUITE]: '#api_key_makersuite',
|
[SECRET_KEYS.MAKERSUITE]: '#api_key_makersuite',
|
||||||
[SECRET_KEYS.APHRODITE]: '#api_key_aphrodite',
|
[SECRET_KEYS.APHRODITE]: '#api_key_aphrodite',
|
||||||
[SECRET_KEYS.TABBY]: '#api_key_tabby',
|
[SECRET_KEYS.TABBY]: '#api_key_tabby',
|
||||||
|
[SECRET_KEYS.MISTRALAI]: '#api_key_mistralai',
|
||||||
};
|
};
|
||||||
|
|
||||||
async function clearSecret() {
|
async function clearSecret() {
|
||||||
|
@ -384,6 +384,10 @@ export function getTokenizerModel() {
|
|||||||
return claudeTokenizer;
|
return claudeTokenizer;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (oai_settings.chat_completion_source == chat_completion_sources.MISTRALAI) {
|
||||||
|
return mistralTokenizer;
|
||||||
|
}
|
||||||
|
|
||||||
// Default to Turbo 3.5
|
// Default to Turbo 3.5
|
||||||
return turboTokenizer;
|
return turboTokenizer;
|
||||||
}
|
}
|
||||||
|
@ -159,6 +159,7 @@ const CHAT_COMPLETION_SOURCES = {
|
|||||||
OPENROUTER: 'openrouter',
|
OPENROUTER: 'openrouter',
|
||||||
AI21: 'ai21',
|
AI21: 'ai21',
|
||||||
MAKERSUITE: 'makersuite',
|
MAKERSUITE: 'makersuite',
|
||||||
|
MISTRALAI: 'mistralai',
|
||||||
};
|
};
|
||||||
|
|
||||||
const UPLOADS_PATH = './uploads';
|
const UPLOADS_PATH = './uploads';
|
||||||
|
@ -392,6 +392,95 @@ async function sendAI21Request(request, response) {
|
|||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sends a request to MistralAI API.
|
||||||
|
* @param {express.Request} request Express request
|
||||||
|
* @param {express.Response} response Express response
|
||||||
|
*/
|
||||||
|
async function sendMistralAIRequest(request, response) {
|
||||||
|
const apiKey = readSecret(SECRET_KEYS.MISTRALAI);
|
||||||
|
|
||||||
|
if (!apiKey) {
|
||||||
|
console.log('MistralAI API key is missing.');
|
||||||
|
return response.status(400).send({ error: true });
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
//must send a user role as last message
|
||||||
|
const messages = Array.isArray(request.body.messages) ? request.body.messages : [];
|
||||||
|
const lastMsg = messages[messages.length - 1];
|
||||||
|
if (messages.length > 0 && lastMsg && (lastMsg.role === 'system' || lastMsg.role === 'assistant')) {
|
||||||
|
lastMsg.role = 'user';
|
||||||
|
if (lastMsg.role === 'assistant') {
|
||||||
|
lastMsg.content = lastMsg.name + ': ' + lastMsg.content;
|
||||||
|
} else if (lastMsg.role === 'system') {
|
||||||
|
lastMsg.content = '[INST] ' + lastMsg.content + ' [/INST]';
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
//system prompts can be stacked at the start, but any futher sys prompts after the first user/assistant message will break the model
|
||||||
|
let encounteredNonSystemMessage = false;
|
||||||
|
messages.forEach(msg => {
|
||||||
|
if ((msg.role === 'user' || msg.role === 'assistant') && !encounteredNonSystemMessage) {
|
||||||
|
encounteredNonSystemMessage = true;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (encounteredNonSystemMessage && msg.role === 'system') {
|
||||||
|
msg.role = 'user';
|
||||||
|
//unsure if the instruct version is what they've deployed on their endpoints and if this will make a difference or not.
|
||||||
|
//it should be better than just sending the message as a user role without context though
|
||||||
|
msg.content = '[INST] ' + msg.content + ' [/INST]';
|
||||||
|
}
|
||||||
|
});
|
||||||
|
const controller = new AbortController();
|
||||||
|
request.socket.removeAllListeners('close');
|
||||||
|
request.socket.on('close', function () {
|
||||||
|
controller.abort();
|
||||||
|
});
|
||||||
|
|
||||||
|
const config = {
|
||||||
|
method: 'POST',
|
||||||
|
headers: {
|
||||||
|
'Content-Type': 'application/json',
|
||||||
|
'Authorization': 'Bearer ' + apiKey,
|
||||||
|
},
|
||||||
|
body: JSON.stringify({
|
||||||
|
'model': request.body.model,
|
||||||
|
'messages': messages,
|
||||||
|
'temperature': request.body.temperature,
|
||||||
|
'top_p': request.body.top_p,
|
||||||
|
'max_tokens': request.body.max_tokens,
|
||||||
|
'stream': request.body.stream,
|
||||||
|
'safe_mode': request.body.safe_mode,
|
||||||
|
'random_seed': request.body.seed === -1 ? undefined : request.body.seed,
|
||||||
|
}),
|
||||||
|
signal: controller.signal,
|
||||||
|
timeout: 0,
|
||||||
|
};
|
||||||
|
|
||||||
|
const generateResponse = await fetch('https://api.mistral.ai/v1/chat/completions', config);
|
||||||
|
if (request.body.stream) {
|
||||||
|
forwardFetchResponse(generateResponse, response);
|
||||||
|
} else {
|
||||||
|
if (!generateResponse.ok) {
|
||||||
|
console.log(`MistralAI API returned error: ${generateResponse.status} ${generateResponse.statusText} ${await generateResponse.text()}`);
|
||||||
|
// a 401 unauthorized response breaks the frontend auth, so return a 500 instead. prob a better way of dealing with this.
|
||||||
|
// 401s are already handled by the streaming processor and dont pop up an error toast, that should probably be fixed too.
|
||||||
|
return response.status(generateResponse.status === 401 ? 500 : generateResponse.status).send({ error: true });
|
||||||
|
}
|
||||||
|
const generateResponseJson = await generateResponse.json();
|
||||||
|
return response.send(generateResponseJson);
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
console.log('Error communicating with MistralAI API: ', error);
|
||||||
|
if (!response.headersSent) {
|
||||||
|
response.send({ error: true });
|
||||||
|
} else {
|
||||||
|
response.end();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
const router = express.Router();
|
const router = express.Router();
|
||||||
|
|
||||||
router.post('/status', jsonParser, async function (request, response_getstatus_openai) {
|
router.post('/status', jsonParser, async function (request, response_getstatus_openai) {
|
||||||
@ -401,15 +490,21 @@ router.post('/status', jsonParser, async function (request, response_getstatus_o
|
|||||||
let api_key_openai;
|
let api_key_openai;
|
||||||
let headers;
|
let headers;
|
||||||
|
|
||||||
if (request.body.chat_completion_source !== CHAT_COMPLETION_SOURCES.OPENROUTER) {
|
if (request.body.chat_completion_source === CHAT_COMPLETION_SOURCES.OPENAI) {
|
||||||
api_url = new URL(request.body.reverse_proxy || API_OPENAI).toString();
|
api_url = new URL(request.body.reverse_proxy || API_OPENAI).toString();
|
||||||
api_key_openai = request.body.reverse_proxy ? request.body.proxy_password : readSecret(SECRET_KEYS.OPENAI);
|
api_key_openai = request.body.reverse_proxy ? request.body.proxy_password : readSecret(SECRET_KEYS.OPENAI);
|
||||||
headers = {};
|
headers = {};
|
||||||
} else {
|
} else if (request.body.chat_completion_source === CHAT_COMPLETION_SOURCES.OPENROUTER) {
|
||||||
api_url = 'https://openrouter.ai/api/v1';
|
api_url = 'https://openrouter.ai/api/v1';
|
||||||
api_key_openai = readSecret(SECRET_KEYS.OPENROUTER);
|
api_key_openai = readSecret(SECRET_KEYS.OPENROUTER);
|
||||||
// OpenRouter needs to pass the referer: https://openrouter.ai/docs
|
// OpenRouter needs to pass the referer: https://openrouter.ai/docs
|
||||||
headers = { 'HTTP-Referer': request.headers.referer };
|
headers = { 'HTTP-Referer': request.headers.referer };
|
||||||
|
} else if (request.body.chat_completion_source === CHAT_COMPLETION_SOURCES.MISTRALAI) {
|
||||||
|
api_url = 'https://api.mistral.ai/v1';
|
||||||
|
api_key_openai = readSecret(SECRET_KEYS.MISTRALAI);
|
||||||
|
} else {
|
||||||
|
console.log('This chat completion source is not supported yet.');
|
||||||
|
return response_getstatus_openai.status(400).send({ error: true });
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!api_key_openai && !request.body.reverse_proxy) {
|
if (!api_key_openai && !request.body.reverse_proxy) {
|
||||||
@ -444,6 +539,9 @@ router.post('/status', jsonParser, async function (request, response_getstatus_o
|
|||||||
});
|
});
|
||||||
|
|
||||||
console.log('Available OpenRouter models:', models);
|
console.log('Available OpenRouter models:', models);
|
||||||
|
} else if (request.body.chat_completion_source === CHAT_COMPLETION_SOURCES.MISTRALAI) {
|
||||||
|
const models = data?.data;
|
||||||
|
console.log(models);
|
||||||
} else {
|
} else {
|
||||||
const models = data?.data;
|
const models = data?.data;
|
||||||
|
|
||||||
@ -551,6 +649,7 @@ router.post('/generate', jsonParser, function (request, response) {
|
|||||||
case CHAT_COMPLETION_SOURCES.SCALE: return sendScaleRequest(request, response);
|
case CHAT_COMPLETION_SOURCES.SCALE: return sendScaleRequest(request, response);
|
||||||
case CHAT_COMPLETION_SOURCES.AI21: return sendAI21Request(request, response);
|
case CHAT_COMPLETION_SOURCES.AI21: return sendAI21Request(request, response);
|
||||||
case CHAT_COMPLETION_SOURCES.MAKERSUITE: return sendMakerSuiteRequest(request, response);
|
case CHAT_COMPLETION_SOURCES.MAKERSUITE: return sendMakerSuiteRequest(request, response);
|
||||||
|
case CHAT_COMPLETION_SOURCES.MISTRALAI: return sendMistralAIRequest(request, response);
|
||||||
}
|
}
|
||||||
|
|
||||||
let apiUrl;
|
let apiUrl;
|
||||||
|
@ -25,6 +25,7 @@ const SECRET_KEYS = {
|
|||||||
DEEPLX_URL: 'deeplx_url',
|
DEEPLX_URL: 'deeplx_url',
|
||||||
MAKERSUITE: 'api_key_makersuite',
|
MAKERSUITE: 'api_key_makersuite',
|
||||||
SERPAPI: 'api_key_serpapi',
|
SERPAPI: 'api_key_serpapi',
|
||||||
|
MISTRALAI: 'api_key_mistralai',
|
||||||
};
|
};
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -12,8 +12,9 @@ const { jsonParser } = require('../express-common');
|
|||||||
*/
|
*/
|
||||||
async function getVector(source, text) {
|
async function getVector(source, text) {
|
||||||
switch (source) {
|
switch (source) {
|
||||||
|
case 'mistral':
|
||||||
case 'openai':
|
case 'openai':
|
||||||
return require('../openai-vectors').getOpenAIVector(text);
|
return require('../openai-vectors').getOpenAIVector(text, source);
|
||||||
case 'transformers':
|
case 'transformers':
|
||||||
return require('../embedding').getTransformersVector(text);
|
return require('../embedding').getTransformersVector(text);
|
||||||
case 'palm':
|
case 'palm':
|
||||||
|
@ -1,20 +1,42 @@
|
|||||||
const fetch = require('node-fetch').default;
|
const fetch = require('node-fetch').default;
|
||||||
const { SECRET_KEYS, readSecret } = require('./endpoints/secrets');
|
const { SECRET_KEYS, readSecret } = require('./endpoints/secrets');
|
||||||
|
|
||||||
|
const SOURCES = {
|
||||||
|
'mistral': {
|
||||||
|
secretKey: SECRET_KEYS.MISTRAL,
|
||||||
|
url: 'api.mistral.ai',
|
||||||
|
model: 'mistral-embed',
|
||||||
|
},
|
||||||
|
'openai': {
|
||||||
|
secretKey: SECRET_KEYS.OPENAI,
|
||||||
|
url: 'api.openai.com',
|
||||||
|
model: 'text-embedding-ada-002',
|
||||||
|
},
|
||||||
|
};
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Gets the vector for the given text from OpenAI ada model
|
* Gets the vector for the given text from an OpenAI compatible endpoint.
|
||||||
* @param {string} text - The text to get the vector for
|
* @param {string} text - The text to get the vector for
|
||||||
|
* @param {string} source - The source of the vector
|
||||||
* @returns {Promise<number[]>} - The vector for the text
|
* @returns {Promise<number[]>} - The vector for the text
|
||||||
*/
|
*/
|
||||||
async function getOpenAIVector(text) {
|
async function getOpenAIVector(text, source) {
|
||||||
const key = readSecret(SECRET_KEYS.OPENAI);
|
const config = SOURCES[source];
|
||||||
|
|
||||||
if (!key) {
|
if (!config) {
|
||||||
console.log('No OpenAI key found');
|
console.log('Unknown source', source);
|
||||||
throw new Error('No OpenAI key found');
|
throw new Error('Unknown source');
|
||||||
}
|
}
|
||||||
|
|
||||||
const response = await fetch('https://api.openai.com/v1/embeddings', {
|
const key = readSecret(config.secretKey);
|
||||||
|
|
||||||
|
if (!key) {
|
||||||
|
console.log('No API key found');
|
||||||
|
throw new Error('No API key found');
|
||||||
|
}
|
||||||
|
|
||||||
|
const url = config.url;
|
||||||
|
const response = await fetch(`https://${url}/v1/embeddings`, {
|
||||||
method: 'POST',
|
method: 'POST',
|
||||||
headers: {
|
headers: {
|
||||||
'Content-Type': 'application/json',
|
'Content-Type': 'application/json',
|
||||||
@ -22,22 +44,22 @@ async function getOpenAIVector(text) {
|
|||||||
},
|
},
|
||||||
body: JSON.stringify({
|
body: JSON.stringify({
|
||||||
input: text,
|
input: text,
|
||||||
model: 'text-embedding-ada-002',
|
model: config.model,
|
||||||
}),
|
}),
|
||||||
});
|
});
|
||||||
|
|
||||||
if (!response.ok) {
|
if (!response.ok) {
|
||||||
const text = await response.text();
|
const text = await response.text();
|
||||||
console.log('OpenAI request failed', response.statusText, text);
|
console.log('API request failed', response.statusText, text);
|
||||||
throw new Error('OpenAI request failed');
|
throw new Error('API request failed');
|
||||||
}
|
}
|
||||||
|
|
||||||
const data = await response.json();
|
const data = await response.json();
|
||||||
const vector = data?.data[0]?.embedding;
|
const vector = data?.data[0]?.embedding;
|
||||||
|
|
||||||
if (!Array.isArray(vector)) {
|
if (!Array.isArray(vector)) {
|
||||||
console.log('OpenAI response was not an array');
|
console.log('API response was not an array');
|
||||||
throw new Error('OpenAI response was not an array');
|
throw new Error('API response was not an array');
|
||||||
}
|
}
|
||||||
|
|
||||||
return vector;
|
return vector;
|
||||||
|
@ -46,6 +46,7 @@ async function loadPlugins(app, pluginsPath) {
|
|||||||
|
|
||||||
if (fs.statSync(pluginFilePath).isDirectory()) {
|
if (fs.statSync(pluginFilePath).isDirectory()) {
|
||||||
await loadFromDirectory(app, pluginFilePath);
|
await loadFromDirectory(app, pluginFilePath);
|
||||||
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Not a JavaScript file.
|
// Not a JavaScript file.
|
||||||
|
Loading…
x
Reference in New Issue
Block a user