Closes #3233
This commit is contained in:
Cohee
2024-12-29 20:38:13 +02:00
parent 662f0e9c73
commit 4c7d160d41
10 changed files with 165 additions and 5 deletions

View File

@@ -31,6 +31,7 @@ export const tokenizers = {
QWEN2: 15,
COMMAND_R: 16,
NEMO: 17,
DEEPSEEK: 18,
BEST_MATCH: 99,
MANUAL_SELECTION: 411,
};
@@ -46,6 +47,7 @@ export const ENCODE_TOKENIZERS = [
tokenizers.QWEN2,
tokenizers.COMMAND_R,
tokenizers.NEMO,
tokenizers.DEEPSEEK,
// uncomment when NovelAI releases Kayra and Clio weights, lol
//tokenizers.NERD,
//tokenizers.NERD2,
@@ -133,6 +135,11 @@ const TOKENIZER_URLS = {
decode: '/api/tokenizers/nemo/decode',
count: '/api/tokenizers/nemo/encode',
},
[tokenizers.DEEPSEEK]: {
encode: '/api/tokenizers/deepseek/encode',
decode: '/api/tokenizers/deepseek/decode',
count: '/api/tokenizers/deepseek/encode',
},
[tokenizers.API_TEXTGENERATIONWEBUI]: {
encode: '/api/tokenizers/remote/textgenerationwebui/encode',
count: '/api/tokenizers/remote/textgenerationwebui/encode',
@@ -551,12 +558,17 @@ export function getTokenizerModel() {
const qwen2Tokenizer = 'qwen2';
const commandRTokenizer = 'command-r';
const nemoTokenizer = 'nemo';
const deepseekTokenizer = 'deepseek';
// Assuming no one would use it for different models.. right?
if (oai_settings.chat_completion_source == chat_completion_sources.SCALE) {
return gpt4Tokenizer;
}
if (oai_settings.chat_completion_source == chat_completion_sources.DEEPSEEK) {
return deepseekTokenizer;
}
// Select correct tokenizer for WindowAI proxies
if (oai_settings.chat_completion_source == chat_completion_sources.WINDOWAI && oai_settings.windowai_model) {
if (oai_settings.windowai_model.includes('gpt-4')) {
@@ -619,6 +631,9 @@ export function getTokenizerModel() {
else if (oai_settings.openrouter_model.includes('jamba')) {
return jambaTokenizer;
}
else if (oai_settings.openrouter_model.includes('deepseek')) {
return deepseekTokenizer;
}
}
if (oai_settings.chat_completion_source == chat_completion_sources.COHERE) {