diff --git a/public/index.html b/public/index.html index ea1267429..123c4032d 100644 --- a/public/index.html +++ b/public/index.html @@ -3253,6 +3253,10 @@ + + + + diff --git a/public/scripts/openai.js b/public/scripts/openai.js index 39b3970ad..71b59f93a 100644 --- a/public/scripts/openai.js +++ b/public/scripts/openai.js @@ -4387,7 +4387,7 @@ async function onModelChange() { if (oai_settings.max_context_unlocked) { $('#openai_max_context').attr('max', unlocked_max); } - else if (['sonar', 'sonar-reasoning'].includes(oai_settings.perplexity_model)) { + else if (['sonar', 'sonar-reasoning', 'sonar-reasoning-pro', 'r1-1776'].includes(oai_settings.perplexity_model)) { $('#openai_max_context').attr('max', 127000); } else if (['sonar-pro'].includes(oai_settings.perplexity_model)) { diff --git a/public/scripts/tokenizers.js b/public/scripts/tokenizers.js index 9a4995206..c20ae2706 100644 --- a/public/scripts/tokenizers.js +++ b/public/scripts/tokenizers.js @@ -679,7 +679,7 @@ export function getTokenizerModel() { } if (oai_settings.chat_completion_source === chat_completion_sources.PERPLEXITY) { - if (oai_settings.perplexity_model.includes('sonar-reasoning')) { + if (oai_settings.perplexity_model.includes('sonar-reasoning') || oai_settings.perplexity_model.includes('r1-1776')) { return deepseekTokenizer; } if (oai_settings.perplexity_model.includes('llama-3') || oai_settings.perplexity_model.includes('llama3')) {