Added saving of breakmodel values so that it defaults to it on next load
This commit is contained in:
parent
2cf6b6e650
commit
49fc854e55
11
aiserver.py
11
aiserver.py
|
@ -2563,6 +2563,9 @@ def get_message(msg):
|
||||||
elif(msg['cmd'] == 'list_model'):
|
elif(msg['cmd'] == 'list_model'):
|
||||||
sendModelSelection(menu=msg['data'])
|
sendModelSelection(menu=msg['data'])
|
||||||
elif(msg['cmd'] == 'load_model'):
|
elif(msg['cmd'] == 'load_model'):
|
||||||
|
f = open("settings/" + vars.model.replace('/', '_') + ".breakmodel", "w")
|
||||||
|
f.write(msg['gpu_layers'])
|
||||||
|
f.close()
|
||||||
load_model(use_gpu=msg['use_gpu'], key=msg['key'], gpu_layers=msg['gpu_layers'])
|
load_model(use_gpu=msg['use_gpu'], key=msg['key'], gpu_layers=msg['gpu_layers'])
|
||||||
elif(msg['cmd'] == 'selectmodel'):
|
elif(msg['cmd'] == 'selectmodel'):
|
||||||
if msg['data'] in ('NeoCustom', 'GPT2Custom') and 'path' not in msg:
|
if msg['data'] in ('NeoCustom', 'GPT2Custom') and 'path' not in msg:
|
||||||
|
@ -2574,7 +2577,13 @@ def get_message(msg):
|
||||||
else:
|
else:
|
||||||
layers = get_layer_count(vars.model)
|
layers = get_layer_count(vars.model)
|
||||||
if layers is not None:
|
if layers is not None:
|
||||||
emit('from_server', {'cmd': 'show_layer_bar', 'data': layers, 'gpu_count': torch.cuda.device_count()}, broadcast=True)
|
if path.exists("settings/" + vars.model.replace('/', '_') + ".breakmodel"):
|
||||||
|
f = open("settings/" + vars.model.replace('/', '_') + ".breakmodel", "r")
|
||||||
|
breakmodel = f.read().split(",")
|
||||||
|
f.close()
|
||||||
|
else:
|
||||||
|
breakmodel = [layers for i in range(torch.cuda.device_count())]
|
||||||
|
emit('from_server', {'cmd': 'show_layer_bar', 'data': layers, 'gpu_count': torch.cuda.device_count(), 'breakmodel': breakmodel}, broadcast=True)
|
||||||
else:
|
else:
|
||||||
emit('from_server', {'cmd': 'hide_layer_bar'}, broadcast=True)
|
emit('from_server', {'cmd': 'hide_layer_bar'}, broadcast=True)
|
||||||
elif(msg['cmd'] == 'loadselect'):
|
elif(msg['cmd'] == 'loadselect'):
|
||||||
|
|
|
@ -2410,7 +2410,7 @@ $(document).ready(function(){
|
||||||
$("#modellayers").removeClass("hidden");
|
$("#modellayers").removeClass("hidden");
|
||||||
html = "";
|
html = "";
|
||||||
for (let i=0; i < msg.gpu_count; i++) {
|
for (let i=0; i < msg.gpu_count; i++) {
|
||||||
html += "GPU " + i + ": <input type='range' class='form-range airange' min='0' max='"+msg.data+"' step='1' value='"+msg.data+"' id='gpu_layers"+i+"' onchange='update_gpu_layers();'>";
|
html += "GPU " + i + ": <input type='range' class='form-range airange' min='0' max='"+msg.data+"' step='1' value='"+msg.breakmodel[i]+"' id='gpu_layers"+i+"' onchange='update_gpu_layers();'>";
|
||||||
}
|
}
|
||||||
$("#model_layer_bars").html(html);
|
$("#model_layer_bars").html(html);
|
||||||
$("#gpu_layers_max").html(msg.data);
|
$("#gpu_layers_max").html(msg.data);
|
||||||
|
|
Loading…
Reference in New Issue