Commit Graph

23 Commits

Author SHA1 Message Date
9c218455c4 [chore] Run ESLint 2024-04-12 14:22:12 +03:00
7348199f73 Add llama.cpp slots info endpoint (#2070) 2024-04-11 23:53:32 +03:00
becd17d7d2 Add some llama.cpp-specific endpoints 2024-04-10 08:11:59 +00:00
bf787a9445 support dreamgen non-streaming 2024-03-08 08:32:47 +01:00
5c410986a4 Add support for DreamGen API.
API docs: https://dreamgen.com/docs/models/opus/v1
API keys: https://dreamgen.com/account/api-keys

I decided to base this on the text-completion API since it's more
flexible with SillyTavern's prompt formating capabilities.

This also means that custom context and instruct settings are required.

Will add documentation in a followup PR.
2024-03-07 12:25:48 +01:00
29c4334c46 #1859 Set keep_alive for ollama 2024-02-26 21:09:21 +02:00
f962ad5c02 Add OpenRouter as a text completion source 2024-02-25 22:47:07 +02:00
f82740a238 Change Non-streaming Handler 2024-02-22 15:51:11 -06:00
7c12c836f2 Implement Key Filter 2024-02-20 09:40:35 -06:00
90d5fbc182 Fix non streaming 2024-02-19 09:46:56 -06:00
8075e4cd1e Changes 2024-02-16 09:07:06 -06:00
b5887960b6 Merge branch 'release' into InfermaticAI 2024-02-16 08:53:04 -06:00
cb32e28b82 Allow multiple stop strings for TogetherAI 2024-01-26 17:40:20 +02:00
f29f934c6b Progress 2024-01-24 06:59:27 -06:00
9cf6cef0a4 Fix aborting generation on KoboldCpp via Text Completion 2023-12-21 23:14:28 +02:00
b782a8cc03 Add util for trim v1 2023-12-21 22:40:08 +02:00
4b131067e4 Add local multimodal caption sources 2023-12-20 00:45:45 +02:00
6859e4443e Fix ollama chunk wrapper 2023-12-19 19:17:19 +02:00
44318fef22 Fix double logging of non-streamed replies 2023-12-19 16:49:21 +02:00
67dd52c21b #1309 Ollama text completion backend 2023-12-19 16:38:11 +02:00
edd737e8bd #371 Add llama.cpp inference server support 2023-12-18 22:38:28 +02:00
b0d9f14534 Re-add Together as a text completion source 2023-12-17 23:38:03 +02:00
52de5869fe Rename file, add missing fetch 2023-12-13 02:22:35 +02:00