Use correct tokenizers for logit bias for Mistral and Llama models over OpenRouter

This commit is contained in:
Cohee
2023-11-09 01:03:54 +02:00
parent 1f36fe5193
commit 0e89bf90bc
3 changed files with 88 additions and 42 deletions

View File

@ -59,7 +59,7 @@ import {
resetScrollHeight,
stringFormat,
} from "./utils.js";
import { countTokensOpenAI } from "./tokenizers.js";
import { countTokensOpenAI, getTokenizerModel } from "./tokenizers.js";
import { formatInstructModeChat, formatInstructModeExamples, formatInstructModePrompt, formatInstructModeSystemPrompt } from "./instruct-mode.js";
export {
@ -1541,7 +1541,7 @@ async function calculateLogitBias() {
let result = {};
try {
const reply = await fetch(`/openai_bias?model=${oai_settings.openai_model}`, {
const reply = await fetch(`/openai_bias?model=${getTokenizerModel()}`, {
method: 'POST',
headers: getRequestHeaders(),
body,