Cohee
e8985c259c
Merge branch 'EugeoSynthesisThirtyTwo/release' into staging
2024-02-29 11:34:38 +02:00
Cohee
184fd1622f
Limit to ooba only. Exclude from presets
2024-02-29 11:33:47 +02:00
Cohee
d8956d3e17
Merge branch 'release' into staging
2024-02-29 11:24:40 +02:00
gabriel dhimoila
76669ff8bb
add max_tokens_second
2024-02-29 00:55:25 +01:00
Cohee
d024d7c700
Allow max value for per-entry depth
2024-02-27 23:34:07 +02:00
Cohee
6f5dbc2a52
Merge pull request #1866 from SillyTavern/staging
...
Staging
2024-02-27 21:11:16 +02:00
Cohee
0fcb176408
Bump package version
2024-02-27 21:10:19 +02:00
Cohee
31f39e30c2
Merge pull request #1864 from Fyphen1223/release
...
Update Japanese translation
2024-02-27 20:31:26 +02:00
Cohee
5a236fbccb
Merge pull request #1865 from deciare/edit-message-macros
...
Parse macros when updating message
2024-02-27 20:25:17 +02:00
Deciare
2a4b8ac438
Update displayed prompt bias when auto-saving edits.
...
When Auto-save Message Edits is enabled, the prompt bias string
displayed beneath the textarea wasn't being updated.
2024-02-27 05:17:38 -05:00
Deciare
7885f19e86
Perform macro substitution while updating message.
...
This addresses 3 issues:
1. Prompt bias string was not removed from the text of the edited
message.
2. Macro substitition was not performed in the prompt bias string.
3. Macro substitution was not performed in the edited message text.
2024-02-27 05:17:31 -05:00
Fyphen
1898192d37
Update readme-ja_jp.md
2024-02-27 18:00:56 +09:00
Cohee
29c4334c46
#1859 Set keep_alive for ollama
2024-02-26 21:09:21 +02:00
Cohee
73886c9fff
Merge pull request #1863 from kingbased/mistral
...
mistral-large
2024-02-26 20:10:51 +02:00
based
149a65cf62
migrate model name in old presets to new naming scheme
2024-02-27 02:23:07 +10:00
based
617ae7d02c
???
2024-02-27 01:42:22 +10:00
based
c58d0b2b94
subvers
2024-02-27 01:12:17 +10:00
based
e86fd08d0f
update mistral models
2024-02-27 01:02:02 +10:00
Cohee
f962ad5c02
Add OpenRouter as a text completion source
2024-02-25 22:47:07 +02:00
Cohee
9e5505a7d4
Autocomplete for WI automation IDs
2024-02-25 03:54:40 +02:00
Cohee
fc289126fa
Add event type for text completion generation request settings ready
2024-02-24 21:45:33 +02:00
Cohee
d5bf9fc28c
Non-streaming logprobs for Aphrodite
2024-02-24 20:53:23 +02:00
Cohee
d140b8d5be
Parse non-streaming tabby logprobs
2024-02-24 20:10:53 +02:00
Cohee
3cedf64f66
Add autocomplete for WI inclusion groups
2024-02-24 19:04:44 +02:00
Cohee
0e357c191b
Align label margins
2024-02-24 18:23:58 +02:00
Cohee
3441667336
#1853 Add WI/Script link by entry automation id
2024-02-24 17:22:51 +02:00
Cohee
7b8ac8f4c4
Properly use vector insert setting
2024-02-24 15:57:26 +02:00
Cohee
16833fc238
Merge pull request #1855 from deciare/llamacpp-sampler-order
...
Sampler order for llama.cpp server backend
2024-02-24 15:45:44 +02:00
Cohee
8848818d67
Fix dynatemp neutralization
2024-02-24 15:32:12 +02:00
Cohee
299bd9d563
Merge branch 'staging' into llamacpp-sampler-order
2024-02-24 15:10:58 +02:00
Cohee
13aebc623a
Merge pull request #1854 from deciare/llamacpp-probs
...
Request and display token probabilities from llama.cpp backend
2024-02-24 15:06:28 +02:00
Cohee
eaadfea639
Extend debounce duration of logprobs renderer
2024-02-24 15:03:57 +02:00
Cohee
9287ff18de
Fix for non-streaming
2024-02-24 14:50:06 +02:00
Cohee
dab9bbb514
Merge pull request #1844 from infermaticAI/InfermaticAI
...
Add InfermaticAI as a text completion source
2024-02-24 14:28:09 +02:00
Deciare
445cbda02f
If token probability is a logarithm it'll be < 0
...
No need to read settings to find out if llama.cpp backend is in use...
2024-02-24 00:13:33 -05:00
Deciare
9eba076ae4
Sampler order for llama.cpp server backend
2024-02-23 23:01:04 -05:00
Deciare
936fbac6c5
Merge remote-tracking branch 'origin/staging' into llamacpp-probs
2024-02-23 17:45:54 -05:00
Cohee
737a0bd3ae
Fix purge extras and mistral vectors
2024-02-23 22:37:00 +02:00
Cohee
9b34ac1bde
Merge pull request #1852 from berbant/staging
...
Display TranslateProvider link
2024-02-23 21:43:59 +02:00
Cohee
cb536a7611
Save a list of safe to export secret keys
2024-02-23 21:41:54 +02:00
Cohee
82c5042bad
Prevent extra loop iterations on buffer init
2024-02-23 21:23:44 +02:00
Cohee
4baefeba68
Extend per-entry scan depth limit, add warnings on overflow
2024-02-23 21:18:40 +02:00
Deciare
344b9eedbc
Request token probabilities from llama.cpp backend
...
llama.cpp server token probabilities are given as values ranging from
0 to 1 instead of as logarithms.
2024-02-23 14:01:46 -05:00
NWilson
f82740a238
Change Non-streaming Handler
2024-02-22 15:51:11 -06:00
berbant
bc2010a762
Update secrets.js
2024-02-22 23:55:57 +04:00
berbant
eb89337f51
Update index.js
2024-02-22 23:49:47 +04:00
Cohee
c9f0d61f19
#1851 Substitute macros in new example chat
2024-02-22 18:45:50 +02:00
NWilson
f569424f3e
Merge branch 'staging' into InfermaticAI
2024-02-22 08:32:10 -06:00
Cohee
beb5e470a2
#1069 Fix hoisting of pristine cards in newest sort
2024-02-22 04:48:46 +02:00
Cohee
ece3b2a7c1
Fix Chat Completions status check on settings loading if another API is selected
2024-02-22 04:36:06 +02:00