81251b073a
Implement downloadable tokenizers
...
Closes #2574 , #2754
2024-09-06 16:28:34 +00:00
5fc16a2474
New AI21 Jamba + tokenizer
2024-08-26 12:07:36 +03:00
62fed2ff6a
fix: pass model parameter for aphrodite tokenizer
2024-08-24 22:43:20 +00:00
e707def7dd
Add tokenizer for Gemma/Gemini
2024-08-16 00:00:43 +03:00
bc7ac12317
New ChatGPT model
2024-08-14 20:31:25 +03:00
6c87dfe10d
Add a token counting endpoint for vLLM
2024-06-26 13:34:02 +00:00
33b22bd4f8
Use GPT-4o tokenizer for Gemini
...
Use GPT-4o tokenizer for Gemini, as Gemini tokenizer is more similar to GPT-4o's
2024-05-25 13:26:11 +08:00
66454bb711
Add reverse proxy support to Google MakerSuite to allow some Google MakerSuite URLs to no longer be hardcoded with domain names. ( #2307 )
...
* Add reverse proxy support to Google MakerSuite.
* Remove hardcoded URLs for some Google MakerSuite API calls.
* Don't send real key to alt.endpoint
* Fix for image captioning
* Fix key validation
* +fix key check for mistral
* Fix caption key validation
* Fix tokenization endpoint use
---------
Co-authored-by: Cohee <18619528+Cohee1207@users.noreply.github.com >
2024-05-24 21:38:29 +03:00
3113109f0a
Use a proper tokenizer for GPT-4o
2024-05-14 15:30:11 +03:00
7bfd666321
Add llama 3 tokenizer
2024-05-03 23:59:39 +03:00
2bd239fe81
Initial vLLM support
2024-05-02 22:40:40 +00:00
212e61d2a1
Lazy initialization of Claude tokenizer. Add JSDoc for tokenizer handlers
2024-04-26 15:17:02 +03:00
31ba3cf039
Merge branch 'staging' into neo-server
2024-04-11 21:47:10 +03:00
34bca69950
system prompt support for gemini 1.5
2024-04-11 16:38:20 +10:00
b07a6a9a78
Update all endpoints to use user directories
2024-04-07 01:47:07 +03:00
30c52b5b27
Move prompt-converters.js 1 level up
2024-03-20 01:18:51 +02:00
bd223486de
Include additional headers for all supported Text Completion types.
2024-03-14 00:48:08 +02:00
04bb882e90
implement messages api and selector for nuclaude models
2024-03-05 04:40:19 +10:00
10d78f9a25
Aphrodite tokenizer
2024-02-02 01:09:03 +02:00
60044c18a4
Implement Token Probabilities UI using logprobs
2024-01-25 18:34:46 -06:00
edd737e8bd
#371 Add llama.cpp inference server support
2023-12-18 22:38:28 +02:00
698850b514
Merge remote-tracking branch 'fork/staging' into gemini
...
# Conflicts:
# server.js
# src/endpoints/prompt-converters.js
# src/endpoints/tokenizers.js
2023-12-15 01:35:17 +10:00
2a5340232d
Move prompt converters to a separate module. Camelcase local variables and add missing JSDocs.
2023-12-14 16:00:17 +02:00
348253fd98
Fix import path
2023-12-14 15:36:44 +02:00
3e82a7d439
tokenizer changes and fixes. + a toggle
2023-12-14 16:31:08 +10:00
e26159c00d
refactor and rework palm request to work with the 'content' format and added an endpoint for googles tokenizer
2023-12-14 15:49:50 +10:00
dba66e756a
Move chat completions API endpoints to module
2023-12-13 18:53:22 -05:00
cebd6e9e0f
Add API token ids from KoboldCpp
2023-12-14 01:28:18 +02:00
30502ac949
Split up Kobold and textgenerationwebui endpoints
...
The endpoint was one big if/else statement that did two entirely
different things depending on the value of main_api. It makes more sense
for those to be two separate endpoints.
2023-12-09 20:26:24 -05:00
04e92efe29
Move API tokenization endpoint into /tokenizers
...
Requires extracting some more functions out of server.js.
2023-12-09 19:35:49 -05:00
4e073250a2
Use Express router for tokenizers endpoint
2023-12-04 13:00:13 -05:00
57bc95133e
Rename tokenizer routes
...
They're all under tokenizers/ now, and there are "count", "encode", and
"decode" endpoints. This forms a clearer hierarchy.
2023-12-04 10:17:43 -05:00
c073a89e2c
Move endpoints into their own folder
2023-12-04 10:17:43 -05:00