Commit Graph

3015 Commits

Author SHA1 Message Date
Juha Jeronen f8f6dd1d90 Manage chat files: fragment search a.k.a. swoop 2024-02-29 14:22:35 +02:00
Cohee 3d84ae026d Fix formatting 2024-02-29 11:50:41 +02:00
Cohee 8981346360
Merge pull request #1861 from berbant/staging
Deleting the current chat when creating a new one
2024-02-29 11:47:05 +02:00
Cohee e8985c259c Merge branch 'EugeoSynthesisThirtyTwo/release' into staging 2024-02-29 11:34:38 +02:00
Cohee 184fd1622f Limit to ooba only. Exclude from presets 2024-02-29 11:33:47 +02:00
gabriel dhimoila 76669ff8bb add max_tokens_second 2024-02-29 00:55:25 +01:00
berbant a85a2bbab1
Merge branch 'SillyTavern:staging' into staging 2024-02-28 22:46:43 +04:00
Cohee d024d7c700 Allow max value for per-entry depth 2024-02-27 23:34:07 +02:00
Deciare 2a4b8ac438 Update displayed prompt bias when auto-saving edits.
When Auto-save Message Edits is enabled, the prompt bias string
displayed beneath the textarea wasn't being updated.
2024-02-27 05:17:38 -05:00
Deciare 7885f19e86 Perform macro substitution while updating message.
This addresses 3 issues:
1. Prompt bias string was not removed from the text of the edited
  message.
2. Macro substitition was not performed in the prompt bias string.
3. Macro substitution was not performed in the edited message text.
2024-02-27 05:17:31 -05:00
based 149a65cf62 migrate model name in old presets to new naming scheme 2024-02-27 02:23:07 +10:00
based c58d0b2b94 subvers 2024-02-27 01:12:17 +10:00
based e86fd08d0f update mistral models 2024-02-27 01:02:02 +10:00
Cohee f962ad5c02 Add OpenRouter as a text completion source 2024-02-25 22:47:07 +02:00
berbant 3c620effaf
Update script.js 2024-02-25 21:19:28 +04:00
berbant 670f08fad2
Update group-chats.js
After deleting a group chat, the oldest chat became active. I've fixed it so that the most recent chat becomes active instead.
2024-02-25 21:11:56 +04:00
Cohee 9e5505a7d4 Autocomplete for WI automation IDs 2024-02-25 03:54:40 +02:00
Cohee fc289126fa Add event type for text completion generation request settings ready 2024-02-24 21:45:33 +02:00
Cohee d5bf9fc28c Non-streaming logprobs for Aphrodite 2024-02-24 20:53:23 +02:00
Cohee d140b8d5be Parse non-streaming tabby logprobs 2024-02-24 20:10:53 +02:00
Cohee 3cedf64f66 Add autocomplete for WI inclusion groups 2024-02-24 19:04:44 +02:00
Cohee 0e357c191b Align label margins 2024-02-24 18:23:58 +02:00
Cohee 3441667336 #1853 Add WI/Script link by entry automation id 2024-02-24 17:22:51 +02:00
Cohee 7b8ac8f4c4 Properly use vector insert setting 2024-02-24 15:57:26 +02:00
Cohee 8848818d67 Fix dynatemp neutralization 2024-02-24 15:32:12 +02:00
Cohee 299bd9d563 Merge branch 'staging' into llamacpp-sampler-order 2024-02-24 15:10:58 +02:00
Cohee 13aebc623a
Merge pull request #1854 from deciare/llamacpp-probs
Request and display token probabilities from llama.cpp backend
2024-02-24 15:06:28 +02:00
Cohee eaadfea639 Extend debounce duration of logprobs renderer 2024-02-24 15:03:57 +02:00
Cohee 9287ff18de Fix for non-streaming 2024-02-24 14:50:06 +02:00
Cohee dab9bbb514
Merge pull request #1844 from infermaticAI/InfermaticAI
Add InfermaticAI as a text completion source
2024-02-24 14:28:09 +02:00
Deciare 445cbda02f If token probability is a logarithm it'll be < 0
No need to read settings to find out if llama.cpp backend is in use...
2024-02-24 00:13:33 -05:00
Deciare 9eba076ae4 Sampler order for llama.cpp server backend 2024-02-23 23:01:04 -05:00
Deciare 936fbac6c5 Merge remote-tracking branch 'origin/staging' into llamacpp-probs 2024-02-23 17:45:54 -05:00
Cohee 737a0bd3ae Fix purge extras and mistral vectors 2024-02-23 22:37:00 +02:00
Cohee 9b34ac1bde
Merge pull request #1852 from berbant/staging
Display TranslateProvider link
2024-02-23 21:43:59 +02:00
Cohee cb536a7611 Save a list of safe to export secret keys 2024-02-23 21:41:54 +02:00
Cohee 82c5042bad Prevent extra loop iterations on buffer init 2024-02-23 21:23:44 +02:00
Cohee 4baefeba68 Extend per-entry scan depth limit, add warnings on overflow 2024-02-23 21:18:40 +02:00
Deciare 344b9eedbc Request token probabilities from llama.cpp backend
llama.cpp server token probabilities are given as values ranging from
0 to 1 instead of as logarithms.
2024-02-23 14:01:46 -05:00
berbant eb89337f51
Update index.js 2024-02-22 23:49:47 +04:00
Cohee c9f0d61f19 #1851 Substitute macros in new example chat 2024-02-22 18:45:50 +02:00
NWilson f569424f3e Merge branch 'staging' into InfermaticAI 2024-02-22 08:32:10 -06:00
Cohee ece3b2a7c1 Fix Chat Completions status check on settings loading if another API is selected 2024-02-22 04:36:06 +02:00
Cohee 0ccdfe4bb7 Fix duped line 2024-02-22 02:45:35 +02:00
Cohee 40aa971d11 Merge branch 'staging' into sampler-order-ooba 2024-02-22 02:44:32 +02:00
Cohee 711fd0517f Merge branch 'staging' into pygimport 2024-02-21 11:26:47 +02:00
Cohee d31195a704 Apply same width for Kobold order
Just in case
2024-02-21 01:02:23 +02:00
Cohee 10fb69f36a Widen the block 2024-02-21 00:59:38 +02:00
Cohee d353fa58d0 Close div properly 2024-02-21 00:56:40 +02:00
Cohee 96f1ce1fce Skill issue? 2024-02-21 00:55:30 +02:00