Commit Graph

3852 Commits

Author SHA1 Message Date
gabriel dhimoila 76669ff8bb add max_tokens_second 2024-02-29 00:55:25 +01:00
Cohee 6f5dbc2a52
Merge pull request #1866 from SillyTavern/staging
Staging
2024-02-27 21:11:16 +02:00
Cohee 0fcb176408 Bump package version 2024-02-27 21:10:19 +02:00
Cohee 31f39e30c2
Merge pull request #1864 from Fyphen1223/release
Update Japanese translation
2024-02-27 20:31:26 +02:00
Cohee 5a236fbccb
Merge pull request #1865 from deciare/edit-message-macros
Parse macros when updating message
2024-02-27 20:25:17 +02:00
Deciare 2a4b8ac438 Update displayed prompt bias when auto-saving edits.
When Auto-save Message Edits is enabled, the prompt bias string
displayed beneath the textarea wasn't being updated.
2024-02-27 05:17:38 -05:00
Deciare 7885f19e86 Perform macro substitution while updating message.
This addresses 3 issues:
1. Prompt bias string was not removed from the text of the edited
  message.
2. Macro substitition was not performed in the prompt bias string.
3. Macro substitution was not performed in the edited message text.
2024-02-27 05:17:31 -05:00
Fyphen 1898192d37
Update readme-ja_jp.md 2024-02-27 18:00:56 +09:00
Cohee 29c4334c46 #1859 Set keep_alive for ollama 2024-02-26 21:09:21 +02:00
Cohee 73886c9fff
Merge pull request #1863 from kingbased/mistral
mistral-large
2024-02-26 20:10:51 +02:00
based 149a65cf62 migrate model name in old presets to new naming scheme 2024-02-27 02:23:07 +10:00
based 617ae7d02c ??? 2024-02-27 01:42:22 +10:00
based c58d0b2b94 subvers 2024-02-27 01:12:17 +10:00
based e86fd08d0f update mistral models 2024-02-27 01:02:02 +10:00
Cohee f962ad5c02 Add OpenRouter as a text completion source 2024-02-25 22:47:07 +02:00
Cohee 9e5505a7d4 Autocomplete for WI automation IDs 2024-02-25 03:54:40 +02:00
Cohee fc289126fa Add event type for text completion generation request settings ready 2024-02-24 21:45:33 +02:00
Cohee d5bf9fc28c Non-streaming logprobs for Aphrodite 2024-02-24 20:53:23 +02:00
Cohee d140b8d5be Parse non-streaming tabby logprobs 2024-02-24 20:10:53 +02:00
Cohee 3cedf64f66 Add autocomplete for WI inclusion groups 2024-02-24 19:04:44 +02:00
Cohee 0e357c191b Align label margins 2024-02-24 18:23:58 +02:00
Cohee 3441667336 #1853 Add WI/Script link by entry automation id 2024-02-24 17:22:51 +02:00
Cohee 7b8ac8f4c4 Properly use vector insert setting 2024-02-24 15:57:26 +02:00
Cohee 16833fc238
Merge pull request #1855 from deciare/llamacpp-sampler-order
Sampler order for llama.cpp server backend
2024-02-24 15:45:44 +02:00
Cohee 8848818d67 Fix dynatemp neutralization 2024-02-24 15:32:12 +02:00
Cohee 299bd9d563 Merge branch 'staging' into llamacpp-sampler-order 2024-02-24 15:10:58 +02:00
Cohee 13aebc623a
Merge pull request #1854 from deciare/llamacpp-probs
Request and display token probabilities from llama.cpp backend
2024-02-24 15:06:28 +02:00
Cohee eaadfea639 Extend debounce duration of logprobs renderer 2024-02-24 15:03:57 +02:00
Cohee 9287ff18de Fix for non-streaming 2024-02-24 14:50:06 +02:00
Cohee dab9bbb514
Merge pull request #1844 from infermaticAI/InfermaticAI
Add InfermaticAI as a text completion source
2024-02-24 14:28:09 +02:00
Deciare 445cbda02f If token probability is a logarithm it'll be < 0
No need to read settings to find out if llama.cpp backend is in use...
2024-02-24 00:13:33 -05:00
Deciare 9eba076ae4 Sampler order for llama.cpp server backend 2024-02-23 23:01:04 -05:00
Deciare 936fbac6c5 Merge remote-tracking branch 'origin/staging' into llamacpp-probs 2024-02-23 17:45:54 -05:00
Cohee 737a0bd3ae Fix purge extras and mistral vectors 2024-02-23 22:37:00 +02:00
Cohee 9b34ac1bde
Merge pull request #1852 from berbant/staging
Display TranslateProvider link
2024-02-23 21:43:59 +02:00
Cohee cb536a7611 Save a list of safe to export secret keys 2024-02-23 21:41:54 +02:00
Cohee 82c5042bad Prevent extra loop iterations on buffer init 2024-02-23 21:23:44 +02:00
Cohee 4baefeba68 Extend per-entry scan depth limit, add warnings on overflow 2024-02-23 21:18:40 +02:00
Deciare 344b9eedbc Request token probabilities from llama.cpp backend
llama.cpp server token probabilities are given as values ranging from
0 to 1 instead of as logarithms.
2024-02-23 14:01:46 -05:00
NWilson f82740a238 Change Non-streaming Handler 2024-02-22 15:51:11 -06:00
berbant bc2010a762
Update secrets.js 2024-02-22 23:55:57 +04:00
berbant eb89337f51
Update index.js 2024-02-22 23:49:47 +04:00
Cohee c9f0d61f19 #1851 Substitute macros in new example chat 2024-02-22 18:45:50 +02:00
NWilson f569424f3e Merge branch 'staging' into InfermaticAI 2024-02-22 08:32:10 -06:00
Cohee beb5e470a2 #1069 Fix hoisting of pristine cards in newest sort 2024-02-22 04:48:46 +02:00
Cohee ece3b2a7c1 Fix Chat Completions status check on settings loading if another API is selected 2024-02-22 04:36:06 +02:00
Cohee 06c3ea7c51
Merge pull request #1811 from kalomaze/sampler-order-ooba
Sampler priority support (for text-generation-webui)
2024-02-22 02:55:38 +02:00
Cohee 0ccdfe4bb7 Fix duped line 2024-02-22 02:45:35 +02:00
Cohee 40aa971d11 Merge branch 'staging' into sampler-order-ooba 2024-02-22 02:44:32 +02:00
Cohee fb6fa54c7f Fix import fetch HTTP method 2024-02-21 19:57:38 +02:00