3008 Commits

Author SHA1 Message Date
johnbenac
d9ff9e796b
Add tip to AllTalk endpoint field. 2024-03-09 15:57:01 -05:00
johnbenac
8036a07df4
clarified placeholder comfyui text 2024-03-06 21:34:31 -05:00
Cohee
1cce26d631 #1891 Fix JWT Scale generation 2024-03-04 22:25:14 +02:00
Deciare
2a4b8ac438 Update displayed prompt bias when auto-saving edits.
When Auto-save Message Edits is enabled, the prompt bias string
displayed beneath the textarea wasn't being updated.
2024-02-27 05:17:38 -05:00
Deciare
7885f19e86 Perform macro substitution while updating message.
This addresses 3 issues:
1. Prompt bias string was not removed from the text of the edited
  message.
2. Macro substitition was not performed in the prompt bias string.
3. Macro substitution was not performed in the edited message text.
2024-02-27 05:17:31 -05:00
based
149a65cf62 migrate model name in old presets to new naming scheme 2024-02-27 02:23:07 +10:00
based
c58d0b2b94 subvers 2024-02-27 01:12:17 +10:00
based
e86fd08d0f update mistral models 2024-02-27 01:02:02 +10:00
Cohee
f962ad5c02 Add OpenRouter as a text completion source 2024-02-25 22:47:07 +02:00
Cohee
9e5505a7d4 Autocomplete for WI automation IDs 2024-02-25 03:54:40 +02:00
Cohee
fc289126fa Add event type for text completion generation request settings ready 2024-02-24 21:45:33 +02:00
Cohee
d5bf9fc28c Non-streaming logprobs for Aphrodite 2024-02-24 20:53:23 +02:00
Cohee
d140b8d5be Parse non-streaming tabby logprobs 2024-02-24 20:10:53 +02:00
Cohee
3cedf64f66 Add autocomplete for WI inclusion groups 2024-02-24 19:04:44 +02:00
Cohee
0e357c191b Align label margins 2024-02-24 18:23:58 +02:00
Cohee
3441667336 #1853 Add WI/Script link by entry automation id 2024-02-24 17:22:51 +02:00
Cohee
7b8ac8f4c4 Properly use vector insert setting 2024-02-24 15:57:26 +02:00
Cohee
8848818d67 Fix dynatemp neutralization 2024-02-24 15:32:12 +02:00
Cohee
299bd9d563 Merge branch 'staging' into llamacpp-sampler-order 2024-02-24 15:10:58 +02:00
Cohee
13aebc623a
Merge pull request #1854 from deciare/llamacpp-probs
Request and display token probabilities from llama.cpp backend
2024-02-24 15:06:28 +02:00
Cohee
eaadfea639 Extend debounce duration of logprobs renderer 2024-02-24 15:03:57 +02:00
Cohee
9287ff18de Fix for non-streaming 2024-02-24 14:50:06 +02:00
Cohee
dab9bbb514
Merge pull request #1844 from infermaticAI/InfermaticAI
Add InfermaticAI as a text completion source
2024-02-24 14:28:09 +02:00
Deciare
445cbda02f If token probability is a logarithm it'll be < 0
No need to read settings to find out if llama.cpp backend is in use...
2024-02-24 00:13:33 -05:00
Deciare
9eba076ae4 Sampler order for llama.cpp server backend 2024-02-23 23:01:04 -05:00
Deciare
936fbac6c5 Merge remote-tracking branch 'origin/staging' into llamacpp-probs 2024-02-23 17:45:54 -05:00
Cohee
737a0bd3ae Fix purge extras and mistral vectors 2024-02-23 22:37:00 +02:00
Cohee
9b34ac1bde
Merge pull request #1852 from berbant/staging
Display TranslateProvider link
2024-02-23 21:43:59 +02:00
Cohee
cb536a7611 Save a list of safe to export secret keys 2024-02-23 21:41:54 +02:00
Cohee
82c5042bad Prevent extra loop iterations on buffer init 2024-02-23 21:23:44 +02:00
Cohee
4baefeba68 Extend per-entry scan depth limit, add warnings on overflow 2024-02-23 21:18:40 +02:00
Deciare
344b9eedbc Request token probabilities from llama.cpp backend
llama.cpp server token probabilities are given as values ranging from
0 to 1 instead of as logarithms.
2024-02-23 14:01:46 -05:00
berbant
eb89337f51
Update index.js 2024-02-22 23:49:47 +04:00
Cohee
c9f0d61f19 #1851 Substitute macros in new example chat 2024-02-22 18:45:50 +02:00
NWilson
f569424f3e Merge branch 'staging' into InfermaticAI 2024-02-22 08:32:10 -06:00
Cohee
ece3b2a7c1 Fix Chat Completions status check on settings loading if another API is selected 2024-02-22 04:36:06 +02:00
Cohee
0ccdfe4bb7 Fix duped line 2024-02-22 02:45:35 +02:00
Cohee
40aa971d11 Merge branch 'staging' into sampler-order-ooba 2024-02-22 02:44:32 +02:00
Cohee
711fd0517f Merge branch 'staging' into pygimport 2024-02-21 11:26:47 +02:00
Cohee
d31195a704 Apply same width for Kobold order
Just in case
2024-02-21 01:02:23 +02:00
Cohee
10fb69f36a Widen the block 2024-02-21 00:59:38 +02:00
Cohee
d353fa58d0 Close div properly 2024-02-21 00:56:40 +02:00
Cohee
96f1ce1fce Skill issue? 2024-02-21 00:55:30 +02:00
Cohee
0c1cf9ff2e Send sampler priority as array 2024-02-21 00:53:54 +02:00
Cohee
f0141b4dd1
Update slash-commands.js 2024-02-20 16:57:00 +02:00
Sneha C
095cd873de
Update slash-commands.js
added the word "persona" to the /sync description to make it easier for users to find.
2024-02-20 16:48:43 +04:00
Cohee
8ba9b5c38b Merge branch 'staging' into sampler-order-ooba 2024-02-20 02:32:33 +02:00
Cohee
8e66a14e37 Add hints to doc strings about additional command prompts 2024-02-20 02:29:14 +02:00
Cohee
79ba026486
Merge pull request #1840 from Wolfsblvt/slash-commands-menu-actions-allow-custom-prompts
Extend impersonate/continue/regenerate with possible custom prompts (via slash commands and popup menu)
2024-02-20 02:26:41 +02:00
kalomaze
cec0698400 Oopsie 2024-02-19 18:24:04 -06:00