mirror of
https://github.com/SillyTavern/SillyTavern.git
synced 2025-06-05 21:59:27 +02:00
Perplexity: Add reasoning model, remove deprecated models
This commit is contained in:
@ -3227,32 +3227,19 @@
|
|||||||
<h4 data-i18n="Perplexity Model">Perplexity Model</h4>
|
<h4 data-i18n="Perplexity Model">Perplexity Model</h4>
|
||||||
<select id="model_perplexity_select">
|
<select id="model_perplexity_select">
|
||||||
<optgroup label="Perplexity Sonar Models">
|
<optgroup label="Perplexity Sonar Models">
|
||||||
|
<option value="sonar">sonar</option>
|
||||||
|
<option value="sonar-pro">sonar-pro</option>
|
||||||
|
<option value="sonar-reasoning">sonar-reasoning</option>
|
||||||
|
</optgroup>
|
||||||
|
<optgroup label="Deprecated Models">
|
||||||
|
<!-- These are scheduled for deprecation after 2/22/2025 -->
|
||||||
<option value="llama-3.1-sonar-small-128k-online">llama-3.1-sonar-small-128k-online</option>
|
<option value="llama-3.1-sonar-small-128k-online">llama-3.1-sonar-small-128k-online</option>
|
||||||
<option value="llama-3.1-sonar-large-128k-online">llama-3.1-sonar-large-128k-online</option>
|
<option value="llama-3.1-sonar-large-128k-online">llama-3.1-sonar-large-128k-online</option>
|
||||||
<option value="llama-3.1-sonar-huge-128k-online">llama-3.1-sonar-huge-128k-online</option>
|
<option value="llama-3.1-sonar-huge-128k-online">llama-3.1-sonar-huge-128k-online</option>
|
||||||
</optgroup>
|
<!-- These are not listed on the site anymore -->
|
||||||
<optgroup label="Perplexity Chat Models">
|
|
||||||
<option value="llama-3.1-sonar-small-128k-chat">llama-3.1-sonar-small-128k-chat</option>
|
<option value="llama-3.1-sonar-small-128k-chat">llama-3.1-sonar-small-128k-chat</option>
|
||||||
<option value="llama-3.1-sonar-large-128k-chat">llama-3.1-sonar-large-128k-chat</option>
|
<option value="llama-3.1-sonar-large-128k-chat">llama-3.1-sonar-large-128k-chat</option>
|
||||||
</optgroup>
|
</optgroup>
|
||||||
<optgroup label="Open-Source Models">
|
|
||||||
<option value="llama-3.1-8b-instruct">llama-3.1-8b-instruct</option>
|
|
||||||
<option value="llama-3.1-70b-instruct">llama-3.1-70b-instruct</option>
|
|
||||||
</optgroup>
|
|
||||||
<optgroup label="Deprecated Models">
|
|
||||||
<option value="llama-3-sonar-small-32k-chat">llama-3-sonar-small-32k-chat</option>
|
|
||||||
<option value="llama-3-sonar-small-32k-online">llama-3-sonar-small-32k-online</option>
|
|
||||||
<option value="llama-3-sonar-large-32k-chat">llama-3-sonar-large-32k-chat</option>
|
|
||||||
<option value="llama-3-sonar-large-32k-online">llama-3-sonar-large-32k-online</option>
|
|
||||||
<option value="sonar-small-chat">sonar-small-chat</option>
|
|
||||||
<option value="sonar-small-online">sonar-small-online</option>
|
|
||||||
<option value="sonar-medium-chat">sonar-medium-chat</option>
|
|
||||||
<option value="sonar-medium-online">sonar-medium-online</option>
|
|
||||||
<option value="llama-3-8b-instruct">llama-3-8b-instruct</option>
|
|
||||||
<option value="llama-3-70b-instruct">llama-3-70b-instruct</option>
|
|
||||||
<option value="mistral-7b-instruct">mistral-7b-instruct (v0.2)</option>
|
|
||||||
<option value="mixtral-8x7b-instruct">mixtral-8x7b-instruct</option>
|
|
||||||
</optgroup>
|
|
||||||
</select>
|
</select>
|
||||||
</div>
|
</div>
|
||||||
<form id="cohere_form" data-source="cohere" action="javascript:void(null);" method="post" enctype="multipart/form-data">
|
<form id="cohere_form" data-source="cohere" action="javascript:void(null);" method="post" enctype="multipart/form-data">
|
||||||
|
@ -258,7 +258,7 @@ const default_settings = {
|
|||||||
ai21_model: 'jamba-1.5-large',
|
ai21_model: 'jamba-1.5-large',
|
||||||
mistralai_model: 'mistral-large-latest',
|
mistralai_model: 'mistral-large-latest',
|
||||||
cohere_model: 'command-r-plus',
|
cohere_model: 'command-r-plus',
|
||||||
perplexity_model: 'llama-3.1-70b-instruct',
|
perplexity_model: 'sonar-pro',
|
||||||
groq_model: 'llama-3.1-70b-versatile',
|
groq_model: 'llama-3.1-70b-versatile',
|
||||||
nanogpt_model: 'gpt-4o-mini',
|
nanogpt_model: 'gpt-4o-mini',
|
||||||
zerooneai_model: 'yi-large',
|
zerooneai_model: 'yi-large',
|
||||||
@ -337,7 +337,7 @@ const oai_settings = {
|
|||||||
ai21_model: 'jamba-1.5-large',
|
ai21_model: 'jamba-1.5-large',
|
||||||
mistralai_model: 'mistral-large-latest',
|
mistralai_model: 'mistral-large-latest',
|
||||||
cohere_model: 'command-r-plus',
|
cohere_model: 'command-r-plus',
|
||||||
perplexity_model: 'llama-3.1-70b-instruct',
|
perplexity_model: 'sonar-pro',
|
||||||
groq_model: 'llama-3.1-70b-versatile',
|
groq_model: 'llama-3.1-70b-versatile',
|
||||||
nanogpt_model: 'gpt-4o-mini',
|
nanogpt_model: 'gpt-4o-mini',
|
||||||
zerooneai_model: 'yi-large',
|
zerooneai_model: 'yi-large',
|
||||||
@ -4380,28 +4380,19 @@ async function onModelChange() {
|
|||||||
if (oai_settings.max_context_unlocked) {
|
if (oai_settings.max_context_unlocked) {
|
||||||
$('#openai_max_context').attr('max', unlocked_max);
|
$('#openai_max_context').attr('max', unlocked_max);
|
||||||
}
|
}
|
||||||
|
else if (['sonar', 'sonar-reasoning'].includes(oai_settings.perplexity_model)) {
|
||||||
|
$('#openai_max_context').attr('max', 127000);
|
||||||
|
}
|
||||||
|
else if (['sonar-pro'].includes(oai_settings.perplexity_model)) {
|
||||||
|
$('#openai_max_context').attr('max', 200000);
|
||||||
|
}
|
||||||
else if (oai_settings.perplexity_model.includes('llama-3.1')) {
|
else if (oai_settings.perplexity_model.includes('llama-3.1')) {
|
||||||
const isOnline = oai_settings.perplexity_model.includes('online');
|
const isOnline = oai_settings.perplexity_model.includes('online');
|
||||||
const contextSize = isOnline ? 128 * 1024 - 4000 : 128 * 1024;
|
const contextSize = isOnline ? 128 * 1024 - 4000 : 128 * 1024;
|
||||||
$('#openai_max_context').attr('max', contextSize);
|
$('#openai_max_context').attr('max', contextSize);
|
||||||
}
|
}
|
||||||
else if (['llama-3-sonar-small-32k-chat', 'llama-3-sonar-large-32k-chat'].includes(oai_settings.perplexity_model)) {
|
|
||||||
$('#openai_max_context').attr('max', max_32k);
|
|
||||||
}
|
|
||||||
else if (['llama-3-sonar-small-32k-online', 'llama-3-sonar-large-32k-online'].includes(oai_settings.perplexity_model)) {
|
|
||||||
$('#openai_max_context').attr('max', 28000);
|
|
||||||
}
|
|
||||||
else if (['sonar-small-chat', 'sonar-medium-chat', 'codellama-70b-instruct', 'mistral-7b-instruct', 'mixtral-8x7b-instruct', 'mixtral-8x22b-instruct'].includes(oai_settings.perplexity_model)) {
|
|
||||||
$('#openai_max_context').attr('max', max_16k);
|
|
||||||
}
|
|
||||||
else if (['llama-3-8b-instruct', 'llama-3-70b-instruct'].includes(oai_settings.perplexity_model)) {
|
|
||||||
$('#openai_max_context').attr('max', max_8k);
|
|
||||||
}
|
|
||||||
else if (['sonar-small-online', 'sonar-medium-online'].includes(oai_settings.perplexity_model)) {
|
|
||||||
$('#openai_max_context').attr('max', 12000);
|
|
||||||
}
|
|
||||||
else {
|
else {
|
||||||
$('#openai_max_context').attr('max', max_4k);
|
$('#openai_max_context').attr('max', max_128k);
|
||||||
}
|
}
|
||||||
oai_settings.openai_max_context = Math.min(Number($('#openai_max_context').attr('max')), oai_settings.openai_max_context);
|
oai_settings.openai_max_context = Math.min(Number($('#openai_max_context').attr('max')), oai_settings.openai_max_context);
|
||||||
$('#openai_max_context').val(oai_settings.openai_max_context).trigger('input');
|
$('#openai_max_context').val(oai_settings.openai_max_context).trigger('input');
|
||||||
|
@ -679,6 +679,9 @@ export function getTokenizerModel() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (oai_settings.chat_completion_source === chat_completion_sources.PERPLEXITY) {
|
if (oai_settings.chat_completion_source === chat_completion_sources.PERPLEXITY) {
|
||||||
|
if (oai_settings.perplexity_model.includes('sonar-reasoning')) {
|
||||||
|
return deepseekTokenizer;
|
||||||
|
}
|
||||||
if (oai_settings.perplexity_model.includes('llama-3') || oai_settings.perplexity_model.includes('llama3')) {
|
if (oai_settings.perplexity_model.includes('llama-3') || oai_settings.perplexity_model.includes('llama3')) {
|
||||||
return llama3Tokenizer;
|
return llama3Tokenizer;
|
||||||
}
|
}
|
||||||
|
Reference in New Issue
Block a user