9c218455c4
[chore] Run ESLint
2024-04-12 14:22:12 +03:00
7348199f73
Add llama.cpp slots info endpoint ( #2070 )
2024-04-11 23:53:32 +03:00
becd17d7d2
Add some llama.cpp-specific endpoints
2024-04-10 08:11:59 +00:00
bf787a9445
support dreamgen non-streaming
2024-03-08 08:32:47 +01:00
5c410986a4
Add support for DreamGen API.
...
API docs: https://dreamgen.com/docs/models/opus/v1
API keys: https://dreamgen.com/account/api-keys
I decided to base this on the text-completion API since it's more
flexible with SillyTavern's prompt formating capabilities.
This also means that custom context and instruct settings are required.
Will add documentation in a followup PR.
2024-03-07 12:25:48 +01:00
29c4334c46
#1859 Set keep_alive for ollama
2024-02-26 21:09:21 +02:00
f962ad5c02
Add OpenRouter as a text completion source
2024-02-25 22:47:07 +02:00
f82740a238
Change Non-streaming Handler
2024-02-22 15:51:11 -06:00
7c12c836f2
Implement Key Filter
2024-02-20 09:40:35 -06:00
90d5fbc182
Fix non streaming
2024-02-19 09:46:56 -06:00
8075e4cd1e
Changes
2024-02-16 09:07:06 -06:00
b5887960b6
Merge branch 'release' into InfermaticAI
2024-02-16 08:53:04 -06:00
cb32e28b82
Allow multiple stop strings for TogetherAI
2024-01-26 17:40:20 +02:00
f29f934c6b
Progress
2024-01-24 06:59:27 -06:00
9cf6cef0a4
Fix aborting generation on KoboldCpp via Text Completion
2023-12-21 23:14:28 +02:00
b782a8cc03
Add util for trim v1
2023-12-21 22:40:08 +02:00
4b131067e4
Add local multimodal caption sources
2023-12-20 00:45:45 +02:00
6859e4443e
Fix ollama chunk wrapper
2023-12-19 19:17:19 +02:00
44318fef22
Fix double logging of non-streamed replies
2023-12-19 16:49:21 +02:00
67dd52c21b
#1309 Ollama text completion backend
2023-12-19 16:38:11 +02:00
edd737e8bd
#371 Add llama.cpp inference server support
2023-12-18 22:38:28 +02:00
b0d9f14534
Re-add Together as a text completion source
2023-12-17 23:38:03 +02:00
52de5869fe
Rename file, add missing fetch
2023-12-13 02:22:35 +02:00