Fix top-k

This commit is contained in:
somebody
2022-12-16 19:56:36 -06:00
parent 802bef8c37
commit 9e6dcc7d6a

View File

@@ -643,7 +643,7 @@ class model_settings(settings):
'koboldai_vars', 'welcome', 'welcome_default', 'simple_randomness', 'simple_creativity', 'simple_repitition', 'koboldai_vars', 'welcome', 'welcome_default', 'simple_randomness', 'simple_creativity', 'simple_repitition',
'badwordsids', 'uid_presets'] 'badwordsids', 'uid_presets']
settings_name = "model" settings_name = "model"
default_settings = {"rep_pen" : 1.1, "rep_pen_slope": 0.7, "rep_pen_range": 1024, "temp": 0.5, "top_p": 0.9, "top_k": 0.0, "top_a": 0.0, "tfs": 1.0, "typical": 1.0, default_settings = {"rep_pen" : 1.1, "rep_pen_slope": 0.7, "rep_pen_range": 1024, "temp": 0.5, "top_p": 0.9, "top_k": 0, "top_a": 0.0, "tfs": 1.0, "typical": 1.0,
"sampler_order": [6,0,1,2,3,4,5]} "sampler_order": [6,0,1,2,3,4,5]}
def __init__(self, socketio, koboldai_vars): def __init__(self, socketio, koboldai_vars):
self.socketio = socketio self.socketio = socketio