Commit Graph

33 Commits

Author SHA1 Message Date
81251b073a Implement downloadable tokenizers
Closes #2574, #2754
2024-09-06 16:28:34 +00:00
5fc16a2474 New AI21 Jamba + tokenizer 2024-08-26 12:07:36 +03:00
62fed2ff6a fix: pass model parameter for aphrodite tokenizer 2024-08-24 22:43:20 +00:00
e707def7dd Add tokenizer for Gemma/Gemini 2024-08-16 00:00:43 +03:00
bc7ac12317 New ChatGPT model 2024-08-14 20:31:25 +03:00
6c87dfe10d Add a token counting endpoint for vLLM 2024-06-26 13:34:02 +00:00
33b22bd4f8 Use GPT-4o tokenizer for Gemini
Use GPT-4o tokenizer for Gemini, as Gemini tokenizer is more similar to GPT-4o's
2024-05-25 13:26:11 +08:00
66454bb711 Add reverse proxy support to Google MakerSuite to allow some Google MakerSuite URLs to no longer be hardcoded with domain names. (#2307)
* Add reverse proxy support to Google MakerSuite.

* Remove hardcoded URLs for some Google MakerSuite API calls.

* Don't send real key to alt.endpoint

* Fix for image captioning

* Fix key validation

* +fix key check for mistral

* Fix caption key validation

* Fix tokenization endpoint use

---------

Co-authored-by: Cohee <18619528+Cohee1207@users.noreply.github.com>
2024-05-24 21:38:29 +03:00
3113109f0a Use a proper tokenizer for GPT-4o 2024-05-14 15:30:11 +03:00
7bfd666321 Add llama 3 tokenizer 2024-05-03 23:59:39 +03:00
2bd239fe81 Initial vLLM support 2024-05-02 22:40:40 +00:00
212e61d2a1 Lazy initialization of Claude tokenizer. Add JSDoc for tokenizer handlers 2024-04-26 15:17:02 +03:00
31ba3cf039 Merge branch 'staging' into neo-server 2024-04-11 21:47:10 +03:00
34bca69950 system prompt support for gemini 1.5 2024-04-11 16:38:20 +10:00
b07a6a9a78 Update all endpoints to use user directories 2024-04-07 01:47:07 +03:00
30c52b5b27 Move prompt-converters.js 1 level up 2024-03-20 01:18:51 +02:00
bd223486de Include additional headers for all supported Text Completion types. 2024-03-14 00:48:08 +02:00
04bb882e90 implement messages api and selector for nuclaude models 2024-03-05 04:40:19 +10:00
10d78f9a25 Aphrodite tokenizer 2024-02-02 01:09:03 +02:00
60044c18a4 Implement Token Probabilities UI using logprobs 2024-01-25 18:34:46 -06:00
edd737e8bd #371 Add llama.cpp inference server support 2023-12-18 22:38:28 +02:00
698850b514 Merge remote-tracking branch 'fork/staging' into gemini
# Conflicts:
#	server.js
#	src/endpoints/prompt-converters.js
#	src/endpoints/tokenizers.js
2023-12-15 01:35:17 +10:00
2a5340232d Move prompt converters to a separate module. Camelcase local variables and add missing JSDocs. 2023-12-14 16:00:17 +02:00
348253fd98 Fix import path 2023-12-14 15:36:44 +02:00
3e82a7d439 tokenizer changes and fixes. + a toggle 2023-12-14 16:31:08 +10:00
e26159c00d refactor and rework palm request to work with the 'content' format and added an endpoint for googles tokenizer 2023-12-14 15:49:50 +10:00
dba66e756a Move chat completions API endpoints to module 2023-12-13 18:53:22 -05:00
cebd6e9e0f Add API token ids from KoboldCpp 2023-12-14 01:28:18 +02:00
30502ac949 Split up Kobold and textgenerationwebui endpoints
The endpoint was one big if/else statement that did two entirely
different things depending on the value of main_api. It makes more sense
for those to be two separate endpoints.
2023-12-09 20:26:24 -05:00
04e92efe29 Move API tokenization endpoint into /tokenizers
Requires extracting some more functions out of server.js.
2023-12-09 19:35:49 -05:00
4e073250a2 Use Express router for tokenizers endpoint 2023-12-04 13:00:13 -05:00
57bc95133e Rename tokenizer routes
They're all under tokenizers/ now, and there are "count", "encode", and
"decode" endpoints. This forms a clearer hierarchy.
2023-12-04 10:17:43 -05:00
c073a89e2c Move endpoints into their own folder 2023-12-04 10:17:43 -05:00