* Kobold Presets fixed
* Help texts fixed
* Scale API for connectAPISlash
* Quick Reply checkboxes fixed
* New Instruct Mode Presets
* More date/time macros
* ChatML context template and instruct prompt format
* Mistral context template and instruct prompt format
* Removed use_default_badwordsids from kobold presets
* Renamed ChatML to Mistral-OpenOrca (ChatML)
* Renamed Mistral-OpenOrca (removed ChatML)
* Removed single_line from kobold presets
* Removed obsolete use_stop_sequence setting
* Ban EOS Token off by default
* Split AI Resp. Conf. in global and preset-specific settings
* Resolve conflicts
* Fix title
* Add translations for new help texts
* Fix i18n.json whitespace
* Make Mistral-OpenOrca system prompt more generic
* Renamed "Mistral-OpenOrca" to "ChatML" again
* More (UI) fixes and improvements
* Sendas hint fixed
---------
Co-authored-by: Cohee <18619528+Cohee1207@users.noreply.github.com>
* The popular mirostat "Gold/Silver/Bronze" community settings
These settings have been popular in the community and are good
candidates for defaults.
* Update Miro Bronze.json
* Update Miro Gold.json
* Update Miro Bronze.json
* Update Miro Silver.json
* Rename Kobold presets
* Update textgen presets
---------
Co-authored-by: Cohee <18619528+Cohee1207@users.noreply.github.com>
After extensive testing, I've adjusted repetition penalty slightly to be the same as simple-proxy-for-tavern's default preset and ooba's LLaMA-Precise settings preset. This fixed some models talking/acting as User.
Here are two presets I've found very useful for Llama 2-based models:
- Deterministic takes away the randomness and is good for testing/comparing models because same input equals same output.
- Storywriter-Llama2 is the Storywriter preset adjusted for Llama 2's 4K context size. It also works well against Llama 2's repetition/looping issues.