Commit Graph

62 Commits

Author SHA1 Message Date
Cohee
3b9fcfae60 Console log Ollama model pulls 2025-03-05 21:42:56 +02:00
Cohee
861c502e44 Return type casts where they were 2025-02-25 22:12:22 +02:00
Cohee
3f03936125 Add config value type converters for numbers and booleans 2025-02-20 21:53:48 +02:00
Cohee
dd7391caaf Ollama: Add num_batch config value 2025-02-13 20:17:33 +02:00
Eradev
055a6527fc PR fixes 2025-02-02 03:10:32 -05:00
Eradev
1f9fa74786 Rewritten minLogLevel feature 2025-01-15 10:02:32 -05:00
Cohee
3167019faf Add generic text completion API type (100% OAI compatible) 2024-12-13 01:12:10 +02:00
Cohee
a86735d743 Fix sendless statuses 2024-11-30 19:30:14 +02:00
Cohee
803b7fca89 Remove custom header 2024-11-23 17:50:32 +02:00
Cohee
362bdf0c46 Don't log when no endpoint available. Remove pointless header 2024-11-23 17:32:12 +02:00
Cohee
049ae541aa Add 'node:' prefix to import 2024-11-23 16:03:13 +02:00
Karl-Johan Alm
c30dde8471 rename /chat_template to /props, and switch to non-trimmed hashes 2024-11-21 11:38:57 +09:00
Karl-Johan Alm
4214c9d894 trim chat template before hashing 2024-11-20 15:12:42 +09:00
Karl-Johan Alm
bb062f5ec9 update endpoint to reflect koboldcpp update 2024-11-20 13:11:51 +09:00
Karl-Johan Alm
feb1b91619 template derivation: add support for llama.cpp server backend 2024-11-19 23:38:38 +09:00
Karl-Johan Alm
cdc0147490 fix error console.log message 2024-11-19 21:41:57 +09:00
Karl-Johan Alm
c2eaae3d42 switch to crypto lib 2024-11-19 21:39:35 +09:00
Karl-Johan Alm
f25ea9f6d6 template derivation: move hash part to backend 2024-11-19 20:09:29 +09:00
Karl-Johan Alm
0e2fdf37a8 feature: derived templates
This PR adds a simple hash based method for picking context and instruct templates based on the chat template, when provided by the back end.
2024-11-19 17:27:25 +09:00
Cohee
24884b3206 TextGenWebUI: remove legacy API mode 2024-10-15 00:32:56 +03:00
Cohee
4fcad0752f [chore] Fix type errors 2024-10-11 21:33:36 +03:00
Cohee
d52b4fbbde Node: Migrate to ES Modules 2024-10-10 22:37:22 +03:00
Cohee
93bf87b035 Shut tokenization errors if not using ooba
Closes #2849
2024-09-22 17:44:19 +03:00
Cohee
2bc87a401e Try detect LM Studio in status check
Fixes #2849
2024-09-15 08:24:28 +03:00
Wolfsblvt
f5b68893d0 Improve error handling of /genraw and /gen
- /generate TC backend returns more status/error texts
- Fix /genraw and /gen returning null/undefined
- Logging errors on /genraw if backend throws an error
- Fixes #2836
2024-09-12 22:41:53 +02:00
Cohee
f305ba7ce7 Configurable ollama keep_alive
Closes #1859
2024-08-11 17:32:31 +03:00
Cohee
5f2a73ac9f Expose "Allow fallback providers" for OpenRouter 2024-07-19 23:34:16 +03:00
Cohee
f7e2d72ac7 Fix eslint 2024-07-07 22:40:03 +03:00
kingbri
0672c8422e Textgen: Add permissions check for TabbyAPI keys
There's no formal permissions checking in ST's UI, so add a temporary
check in the server endpoint before requesting a download.

Signed-off-by: kingbri <bdashore3@proton.me>
2024-07-07 14:40:24 -04:00
kingbri
69077f6a6e Textgen: Add model downloading for TabbyAPI
Tabby has a HuggingFace downloader API endpoint. Add direct support
in SillyTavern by using the same method as ollama, but with the correct
parameters exposed.

Signed-off-by: kingbri <bdashore3@proton.me>
2024-07-07 14:04:04 -04:00
Cohee
4e33253a91 Code clean-up 2024-06-28 19:12:56 +03:00
Cohee
b62cbdeebd Merge branch 'staging' into DarokCx/release 2024-06-28 19:09:12 +03:00
Cohee
bbb1a6e578 Add huggingface inference as text completion source 2024-06-28 18:17:27 +03:00
DarokCx
8608bc92ae no authorization 2024-06-27 10:02:28 -04:00
DarokCx
bd5592de7b Added featherless, connect button not working 2024-06-27 09:06:11 -04:00
Cohee
41ab90bb8e Support more parameters for Infermatic 2024-06-24 19:16:20 +03:00
Cohee
7063fce2af Selectable openrouter providers 2024-05-06 19:26:20 +03:00
sasha0552
2bd239fe81 Initial vLLM support 2024-05-02 22:40:40 +00:00
joenunezb
890cf81627 Fix: InformaticAI response without message in choices 2024-04-23 03:56:50 -07:00
Cohee
9c218455c4 [chore] Run ESLint 2024-04-12 14:22:12 +03:00
sasha0552
7348199f73 Add llama.cpp slots info endpoint (#2070) 2024-04-11 23:53:32 +03:00
sasha0552
becd17d7d2 Add some llama.cpp-specific endpoints 2024-04-10 08:11:59 +00:00
DreamGenX
bf787a9445 support dreamgen non-streaming 2024-03-08 08:32:47 +01:00
DreamGenX
5c410986a4 Add support for DreamGen API.
API docs: https://dreamgen.com/docs/models/opus/v1
API keys: https://dreamgen.com/account/api-keys

I decided to base this on the text-completion API since it's more
flexible with SillyTavern's prompt formating capabilities.

This also means that custom context and instruct settings are required.

Will add documentation in a followup PR.
2024-03-07 12:25:48 +01:00
Cohee
29c4334c46 #1859 Set keep_alive for ollama 2024-02-26 21:09:21 +02:00
Cohee
f962ad5c02 Add OpenRouter as a text completion source 2024-02-25 22:47:07 +02:00
NWilson
f82740a238 Change Non-streaming Handler 2024-02-22 15:51:11 -06:00
NWilson
7c12c836f2 Implement Key Filter 2024-02-20 09:40:35 -06:00
NWilson
90d5fbc182 Fix non streaming 2024-02-19 09:46:56 -06:00
NWilson
8075e4cd1e Changes 2024-02-16 09:07:06 -06:00