* Kobold Presets fixed
* Help texts fixed
* Scale API for connectAPISlash
* Quick Reply checkboxes fixed
* New Instruct Mode Presets
* More date/time macros
* ChatML context template and instruct prompt format
* Mistral context template and instruct prompt format
* Removed use_default_badwordsids from kobold presets
* Renamed ChatML to Mistral-OpenOrca (ChatML)
* Renamed Mistral-OpenOrca (removed ChatML)
* Removed single_line from kobold presets
* Removed obsolete use_stop_sequence setting
* Ban EOS Token off by default
* Split AI Resp. Conf. in global and preset-specific settings
* Resolve conflicts
* Fix title
* Add translations for new help texts
* Fix i18n.json whitespace
* Make Mistral-OpenOrca system prompt more generic
* Renamed "Mistral-OpenOrca" to "ChatML" again
* More (UI) fixes and improvements
* Sendas hint fixed
---------
Co-authored-by: Cohee <18619528+Cohee1207@users.noreply.github.com>
This is a useful feature for those who want to utilize APIs with
proxy middleware for adding extra features or security. For cloud
API safety and abiding by rate limits, this feature only applies to
local backends such as ooba or kobold.
Signed-off-by: kingbri <bdashore3@proton.me>
* Updated some presets and added Cosmic Cube
* Change defaults for NovelAI to select Clio on cold start
* Automatically change the preset to an appropriate default whenever you change the model
* Removed deprecated Top G sampler
* Remove AI Module "Autoselect" and make the auto-instruct work for all modules. This is how NAI is supposed to work.
* Log the response from the API.
* Move the AI Module setting up to the top of the settings window since it isn't part of the preset.
* Refactor phrase_rep_pen to use the actual API strings.
* Clamp the maximum token length to 150 before we call the API.
* Clamp the minimum token length in the UX from 1 to 150.
* Fix bug where the preamble was not initialized on cold start.
* Get rid of extra newline before dinkus.
* Make always_force_name2 default true.