Commit Graph

279 Commits

Author SHA1 Message Date
valadaptive 92bd766bcb Rename chat completions endpoints
OpenAI calls this the "Chat Completions API", in contrast to their
previous "Text Completions API", so that's what I'm naming it; both
because other services besides OpenAI implement it, and to avoid
confusion with the existing /api/openai route used for OpenAI extras.
2023-12-13 18:52:08 -05:00
Cohee 9160de7714 Run macros on impersonation prompt 2023-12-12 19:24:32 +02:00
Cohee 9176f46caf Add /preset command 2023-12-12 19:14:17 +02:00
Cohee b0e7b73a32 Fix streaming processor error handler hooks 2023-12-08 02:01:08 +02:00
valadaptive 5569a63595 Remove legacy_streaming setting
This was a workaround for older versions of Slaude that implemented SSE
improperly. This was fixed in Slaude 7 months ago, so the workaround can
be removed.
2023-12-07 18:00:36 -05:00
valadaptive cdcd913805 Don't stream events if the API returned a 4xx code 2023-12-07 18:00:36 -05:00
valadaptive 5540c165cf Refactor server-sent events parsing
Create one server-sent events stream class which implements the entire
spec (different line endings, chunking, etc) and use it in all the
streaming generators.
2023-12-07 18:00:36 -05:00
Cohee 72adb4c8aa Fix window.ai streaming 2023-12-07 17:42:06 +02:00
Cohee 671df1f62e Fix constant usage 2023-12-04 00:24:23 +02:00
valadaptive e33c8bd955 Replace use_[source] with chat_completion_source
Same as the is_[api] replacement--it's easier to have one enum field
than several mutually-exclusive boolean ones
2023-12-03 15:03:39 -05:00
Cohee 8a1ead531c
Merge pull request #1439 from valadaptive/prompt-manager-class
Convert PromptManagerModule to a class
2023-12-03 21:52:27 +02:00
Cohee 1786b0d340 #1403 Add Aphrodite multi-swipe 2023-12-03 20:40:09 +02:00
valadaptive b8b24540a9 Rename PromptManagerModule to PromptManager
The one place where it was imported renamed it to PromptManager anyway.
2023-12-03 12:14:56 -05:00
Cohee a3bc51bcea Fix type-in max context for OAI 2023-12-03 13:56:22 +02:00
Cohee 64a3564892 lint: Comma dangle 2023-12-02 22:06:57 +02:00
Cohee c63cd87cc0 lint: Require semicolons 2023-12-02 21:11:06 +02:00
valadaptive a37f874e38 Require single quotes 2023-12-02 13:04:51 -05:00
valadaptive 518bb58d5a Enable no-unused-vars lint
This is the big one. Probably needs thorough review to make sure I
didn't accidentally remove any setInterval or fetch calls.
2023-12-02 12:11:19 -05:00
valadaptive c893e2165e Enable no-prototype-builtins lint 2023-12-02 12:10:31 -05:00
valadaptive 0a27275772 Enable no-extra-semi lint 2023-12-02 10:32:26 -05:00
valadaptive 367f3dba27 Enable no-unsafe-finally lint 2023-12-02 10:32:07 -05:00
Cohee 19c6370fa5 Revert preset checkbox update logic 2023-12-01 11:55:05 +02:00
Cohee b96054f337 Update max token limit for palm2 2023-11-30 19:02:31 +02:00
Cohee e9ad55aef2 Add seed input field for OpenAI settings #1412 2023-11-30 02:54:52 +02:00
Cohee d263760b25 #1393 Configurable group nudges, scenario and personality templates for prompt manager 2023-11-27 23:57:56 +02:00
Cohee 61908935f5 Stop string for user-continue. Trim spaces after name2 2023-11-22 16:16:48 +02:00
Cohee 5f77b2f816 Add Claude 2.1 2023-11-21 20:07:37 +02:00
Cohee 73e081dd99 Don't use global state to build Chat Completion prompts 2023-11-21 14:38:15 +02:00
Cohee 0608c0afac Add OpenRouter and Llava to captioning plugin. 2023-11-17 23:19:21 +02:00
Cohee 323b338cdd Add images to quiet prompts if inlining enabled 2023-11-17 01:30:32 +02:00
Cohee d114ebf6fa Add default role for Message class if not set. 2023-11-16 16:20:33 +02:00
Cohee 314aca3f2c Allow disabling system marker prompts 2023-11-14 22:27:07 +02:00
Cohee d3e5f6ebc0 #1343 Move bypass check up 2023-11-12 23:08:24 +02:00
Cohee 9a1d1594d6 Fix formatting in openai.js 2023-11-12 22:14:35 +02:00
artisticMink cc0b4e8174 Access oai_settings instead of dom 2023-11-12 20:55:29 +01:00
artisticMink 3bbbf0d8e4 Put openrouter model sorting in drawer
Renames 'Infinity'k tokens to 'Free'
2023-11-12 19:02:41 +01:00
artisticMink cb2644cdea Add sorting for openrouter models
Alphabetically (default), price, context size
2023-11-12 18:27:56 +01:00
artisticMink a16e34bcef Add optional toggle for grouping openrouter models
By vendor
2023-11-12 15:15:30 +01:00
Cohee 7afe9e6481 #1343 Add status check bypass 2023-11-12 13:23:46 +02:00
Cohee 4c0b3fb7ae Add null checks for OR pricing 2023-11-12 13:07:57 +02:00
Cohee 879502c1e7 Only allow inlining if OAI is the selected API 2023-11-12 00:13:30 +02:00
Cohee 2c4f53e7b5 Add native GPT-4V image inlining 2023-11-12 00:09:48 +02:00
Cohee 2f5e7778cc Don't add items of unknown type to chat completion 2023-11-10 01:08:18 +02:00
Cohee 0e89bf90bc Use correct tokenizers for logit bias for Mistral and Llama models over OpenRouter 2023-11-09 01:03:54 +02:00
Cohee d81354e2a5 Merge branch 'staging' of https://github.com/SillyTavern/SillyTavern into staging 2023-11-08 16:29:02 +02:00
Cohee dbf995fd24 Add character card macros 2023-11-08 16:28:55 +02:00
RossAscends a5fd33d08a Kobold sampler restyle 2023-11-08 23:24:28 +09:00
Cohee 740f6548a2 Increase timeouts of OAI out of quota requests 2023-11-08 12:07:14 +02:00
Cohee b2629d9718 Refactor status checks and indicators. 2023-11-08 02:52:03 +02:00
Cohee 2020d12217 Add new GPT 3.5 turbo model 2023-11-07 00:10:32 +02:00