logger for prompt and gen
This commit is contained in:
parent
66ae5c35c0
commit
d30bbd28a1
70
aiserver.py
70
aiserver.py
|
@ -2849,8 +2849,8 @@ def lua_startup():
|
||||||
except lupa.LuaError as e:
|
except lupa.LuaError as e:
|
||||||
print(colors.RED + "ERROR!" + colors.END)
|
print(colors.RED + "ERROR!" + colors.END)
|
||||||
vars.lua_koboldbridge.obliterate_multiverse()
|
vars.lua_koboldbridge.obliterate_multiverse()
|
||||||
print("{0}{1}{2}".format(colors.RED, "***LUA ERROR***: ", colors.END), end="", file=sys.stderr)
|
logger.debug('LUA ERROR: ' + str(e).replace("\033", ""))
|
||||||
print("{0}{1}{2}".format(colors.RED, str(e).replace("\033", ""), colors.END), file=sys.stderr)
|
logger.warning("Lua engine stopped; please open 'Userscripts' and press Load to reinitialize scripts.")
|
||||||
exit(1)
|
exit(1)
|
||||||
logger.init_ok("LUA bridge", status="OK")
|
logger.init_ok("LUA bridge", status="OK")
|
||||||
|
|
||||||
|
@ -2908,7 +2908,7 @@ def load_lua_scripts():
|
||||||
if(vars.serverstarted):
|
if(vars.serverstarted):
|
||||||
emit('from_server', {'cmd': 'errmsg', 'data': 'Lua script error; please check console.'}, broadcast=True)
|
emit('from_server', {'cmd': 'errmsg', 'data': 'Lua script error; please check console.'}, broadcast=True)
|
||||||
sendUSStatItems()
|
sendUSStatItems()
|
||||||
logger.error('LUA ERROR: ' + str(e).replace("\033", ""))
|
logger.debug('LUA ERROR: ' + str(e).replace("\033", ""))
|
||||||
logger.warning("Lua engine stopped; please open 'Userscripts' and press Load to reinitialize scripts.")
|
logger.warning("Lua engine stopped; please open 'Userscripts' and press Load to reinitialize scripts.")
|
||||||
if(vars.serverstarted):
|
if(vars.serverstarted):
|
||||||
set_aibusy(0)
|
set_aibusy(0)
|
||||||
|
@ -3405,9 +3405,8 @@ def execute_inmod():
|
||||||
vars.lua_running = False
|
vars.lua_running = False
|
||||||
emit('from_server', {'cmd': 'errmsg', 'data': 'Lua script error; please check console.'}, broadcast=True)
|
emit('from_server', {'cmd': 'errmsg', 'data': 'Lua script error; please check console.'}, broadcast=True)
|
||||||
sendUSStatItems()
|
sendUSStatItems()
|
||||||
print("{0}{1}{2}".format(colors.RED, "***LUA ERROR***: ", colors.END), end="", file=sys.stderr)
|
logger.debug('LUA ERROR: ' + str(e).replace("\033", ""))
|
||||||
print("{0}{1}{2}".format(colors.RED, str(e).replace("\033", ""), colors.END), file=sys.stderr)
|
logger.warning("Lua engine stopped; please open 'Userscripts' and press Load to reinitialize scripts.")
|
||||||
print("{0}{1}{2}".format(colors.YELLOW, "Lua engine stopped; please open 'Userscripts' and press Load to reinitialize scripts.", colors.END), file=sys.stderr)
|
|
||||||
set_aibusy(0)
|
set_aibusy(0)
|
||||||
|
|
||||||
def execute_genmod():
|
def execute_genmod():
|
||||||
|
@ -3423,9 +3422,8 @@ def execute_outmod():
|
||||||
vars.lua_running = False
|
vars.lua_running = False
|
||||||
emit('from_server', {'cmd': 'errmsg', 'data': 'Lua script error; please check console.'}, broadcast=True)
|
emit('from_server', {'cmd': 'errmsg', 'data': 'Lua script error; please check console.'}, broadcast=True)
|
||||||
sendUSStatItems()
|
sendUSStatItems()
|
||||||
print("{0}{1}{2}".format(colors.RED, "***LUA ERROR***: ", colors.END), end="", file=sys.stderr)
|
logger.debug('LUA ERROR: ' + str(e).replace("\033", ""))
|
||||||
print("{0}{1}{2}".format(colors.RED, str(e).replace("\033", ""), colors.END), file=sys.stderr)
|
logger.warning("Lua engine stopped; please open 'Userscripts' and press Load to reinitialize scripts.")
|
||||||
print("{0}{1}{2}".format(colors.YELLOW, "Lua engine stopped; please open 'Userscripts' and press Load to reinitialize scripts.", colors.END), file=sys.stderr)
|
|
||||||
set_aibusy(0)
|
set_aibusy(0)
|
||||||
if(vars.lua_koboldbridge.resend_settings_required):
|
if(vars.lua_koboldbridge.resend_settings_required):
|
||||||
vars.lua_koboldbridge.resend_settings_required = False
|
vars.lua_koboldbridge.resend_settings_required = False
|
||||||
|
@ -3445,7 +3443,7 @@ def execute_outmod():
|
||||||
#==================================================================#
|
#==================================================================#
|
||||||
@socketio.on('connect')
|
@socketio.on('connect')
|
||||||
def do_connect():
|
def do_connect():
|
||||||
print("{0}Client connected!{1}".format(colors.GREEN, colors.END))
|
logger.info("Client connected!")
|
||||||
emit('from_server', {'cmd': 'setchatname', 'data': vars.chatname})
|
emit('from_server', {'cmd': 'setchatname', 'data': vars.chatname})
|
||||||
emit('from_server', {'cmd': 'setanotetemplate', 'data': vars.authornotetemplate})
|
emit('from_server', {'cmd': 'setanotetemplate', 'data': vars.authornotetemplate})
|
||||||
emit('from_server', {'cmd': 'connected', 'smandelete': vars.smandelete, 'smanrename': vars.smanrename, 'modelname': getmodelname()})
|
emit('from_server', {'cmd': 'connected', 'smandelete': vars.smandelete, 'smanrename': vars.smanrename, 'modelname': getmodelname()})
|
||||||
|
@ -3498,7 +3496,7 @@ def do_connect():
|
||||||
@socketio.on('message')
|
@socketio.on('message')
|
||||||
def get_message(msg):
|
def get_message(msg):
|
||||||
if not vars.quiet:
|
if not vars.quiet:
|
||||||
print("{0}Data received:{1}{2}".format(colors.GREEN, msg, colors.END))
|
logger.debug(f"Data received: {msg}")
|
||||||
# Submit action
|
# Submit action
|
||||||
if(msg['cmd'] == 'submit'):
|
if(msg['cmd'] == 'submit'):
|
||||||
if(vars.mode == "play"):
|
if(vars.mode == "play"):
|
||||||
|
@ -3788,8 +3786,7 @@ def get_message(msg):
|
||||||
elif(msg['cmd'] == 'list_model'):
|
elif(msg['cmd'] == 'list_model'):
|
||||||
sendModelSelection(menu=msg['data'])
|
sendModelSelection(menu=msg['data'])
|
||||||
elif(msg['cmd'] == 'load_model'):
|
elif(msg['cmd'] == 'load_model'):
|
||||||
print(msg)
|
logger.debug(vars.model_selected)
|
||||||
print(vars.model_selected)
|
|
||||||
if not os.path.exists("settings/"):
|
if not os.path.exists("settings/"):
|
||||||
os.mkdir("settings")
|
os.mkdir("settings")
|
||||||
changed = True
|
changed = True
|
||||||
|
@ -3823,7 +3820,7 @@ def get_message(msg):
|
||||||
vars.cluster_requested_models = msg['online_model']
|
vars.cluster_requested_models = msg['online_model']
|
||||||
load_model(use_gpu=msg['use_gpu'], gpu_layers=msg['gpu_layers'], disk_layers=msg['disk_layers'], online_model=msg['online_model'])
|
load_model(use_gpu=msg['use_gpu'], gpu_layers=msg['gpu_layers'], disk_layers=msg['disk_layers'], online_model=msg['online_model'])
|
||||||
elif(msg['cmd'] == 'show_model'):
|
elif(msg['cmd'] == 'show_model'):
|
||||||
print("Model Name: {}".format(getmodelname()))
|
logger.info(f"Model Name: {getmodelname()}")
|
||||||
emit('from_server', {'cmd': 'show_model_name', 'data': getmodelname()}, broadcast=True)
|
emit('from_server', {'cmd': 'show_model_name', 'data': getmodelname()}, broadcast=True)
|
||||||
elif(msg['cmd'] == 'selectmodel'):
|
elif(msg['cmd'] == 'selectmodel'):
|
||||||
# This is run when a model line is selected from the UI (line from the model_menu variable) that is tagged as not a menu
|
# This is run when a model line is selected from the UI (line from the model_menu variable) that is tagged as not a menu
|
||||||
|
@ -4229,8 +4226,8 @@ def actionsubmit(data, actionmode=0, force_submit=False, force_prompt_gen=False,
|
||||||
try:
|
try:
|
||||||
alternatives = [item['Text'] for item in vars.actions_metadata[len(vars.actions)-1]["Alternative Text"]]
|
alternatives = [item['Text'] for item in vars.actions_metadata[len(vars.actions)-1]["Alternative Text"]]
|
||||||
except:
|
except:
|
||||||
print(len(vars.actions))
|
logger.debug(len(vars.actions))
|
||||||
print(vars.actions_metadata)
|
logger.debug(vars.actions_metadata)
|
||||||
raise
|
raise
|
||||||
if data in alternatives:
|
if data in alternatives:
|
||||||
alternatives = [item for item in vars.actions_metadata[vars.actions.get_last_key() ]["Alternative Text"] if item['Text'] != data]
|
alternatives = [item for item in vars.actions_metadata[vars.actions.get_last_key() ]["Alternative Text"] if item['Text'] != data]
|
||||||
|
@ -4282,7 +4279,10 @@ def apiactionsubmit_generate(txt, minimum, maximum):
|
||||||
vars.generated_tkns = 0
|
vars.generated_tkns = 0
|
||||||
|
|
||||||
if not vars.quiet:
|
if not vars.quiet:
|
||||||
print("{0}Min:{1}, Max:{2}, Txt:{3}{4}".format(colors.YELLOW, minimum, maximum, utils.decodenewlines(tokenizer.decode(txt)), colors.END))
|
logger.debug(f"Prompt Min:{minimum}, Max:{maximum}")
|
||||||
|
split_prompt = utils.decodenewlines(tokenizer.decode(txt)).split('\n')
|
||||||
|
for line in split_prompt:
|
||||||
|
logger.prompt(line)
|
||||||
|
|
||||||
# Clear CUDA cache if using GPU
|
# Clear CUDA cache if using GPU
|
||||||
if(vars.hascuda and (vars.usegpu or vars.breakmodel)):
|
if(vars.hascuda and (vars.usegpu or vars.breakmodel)):
|
||||||
|
@ -4309,7 +4309,10 @@ def apiactionsubmit_tpumtjgenerate(txt, minimum, maximum):
|
||||||
tpu_mtj_backend.set_rng_seed(vars.seed)
|
tpu_mtj_backend.set_rng_seed(vars.seed)
|
||||||
|
|
||||||
if not vars.quiet:
|
if not vars.quiet:
|
||||||
print("{0}Min:{1}, Max:{2}, Txt:{3}{4}".format(colors.YELLOW, minimum, maximum, utils.decodenewlines(tokenizer.decode(txt)), colors.END))
|
logger.debug(f"Prompt Min:{minimum}, Max:{maximum}")
|
||||||
|
split_prompt = utils.decodenewlines(tokenizer.decode(txt)).split('\n')
|
||||||
|
for line in split_prompt:
|
||||||
|
logger.prompt(line)
|
||||||
|
|
||||||
vars._actions = vars.actions
|
vars._actions = vars.actions
|
||||||
vars._prompt = vars.prompt
|
vars._prompt = vars.prompt
|
||||||
|
@ -4821,7 +4824,10 @@ def generate(txt, minimum, maximum, found_entries=None):
|
||||||
found_entries = tuple(found_entries.copy() for _ in range(vars.numseqs))
|
found_entries = tuple(found_entries.copy() for _ in range(vars.numseqs))
|
||||||
|
|
||||||
if not vars.quiet:
|
if not vars.quiet:
|
||||||
print("{0}Min:{1}, Max:{2}, Txt:{3}{4}".format(colors.YELLOW, minimum, maximum, utils.decodenewlines(tokenizer.decode(txt)), colors.END))
|
logger.debug(f"Prompt Min:{minimum}, Max:{maximum}")
|
||||||
|
split_prompt = utils.decodenewlines(tokenizer.decode(txt)).split('\n')
|
||||||
|
for line in split_prompt:
|
||||||
|
logger.prompt(line)
|
||||||
|
|
||||||
# Store context in memory to use it for comparison with generated content
|
# Store context in memory to use it for comparison with generated content
|
||||||
vars.lastctx = utils.decodenewlines(tokenizer.decode(txt))
|
vars.lastctx = utils.decodenewlines(tokenizer.decode(txt))
|
||||||
|
@ -4840,12 +4846,11 @@ def generate(txt, minimum, maximum, found_entries=None):
|
||||||
vars.lua_running = False
|
vars.lua_running = False
|
||||||
emit('from_server', {'cmd': 'errmsg', 'data': 'Lua script error; please check console.'}, broadcast=True)
|
emit('from_server', {'cmd': 'errmsg', 'data': 'Lua script error; please check console.'}, broadcast=True)
|
||||||
sendUSStatItems()
|
sendUSStatItems()
|
||||||
print("{0}{1}{2}".format(colors.RED, "***LUA ERROR***: ", colors.END), end="", file=sys.stderr)
|
logger.debug('LUA ERROR: ' + str(e).replace("\033", ""))
|
||||||
print("{0}{1}{2}".format(colors.RED, str(e).replace("\033", ""), colors.END), file=sys.stderr)
|
logger.warning("Lua engine stopped; please open 'Userscripts' and press Load to reinitialize scripts.")
|
||||||
print("{0}{1}{2}".format(colors.YELLOW, "Lua engine stopped; please open 'Userscripts' and press Load to reinitialize scripts.", colors.END), file=sys.stderr)
|
|
||||||
else:
|
else:
|
||||||
emit('from_server', {'cmd': 'errmsg', 'data': 'Error occurred during generator call; please check console.'}, broadcast=True)
|
emit('from_server', {'cmd': 'errmsg', 'data': 'Error occurred during generator call; please check console.'}, broadcast=True)
|
||||||
print("{0}{1}{2}".format(colors.RED, traceback.format_exc().replace("\033", ""), colors.END), file=sys.stderr)
|
logger.error(traceback.format_exc().replace("\033", ""))
|
||||||
set_aibusy(0)
|
set_aibusy(0)
|
||||||
return
|
return
|
||||||
|
|
||||||
|
@ -4884,7 +4889,9 @@ def generate(txt, minimum, maximum, found_entries=None):
|
||||||
#==================================================================#
|
#==================================================================#
|
||||||
def genresult(genout, flash=True, ignore_formatting=False):
|
def genresult(genout, flash=True, ignore_formatting=False):
|
||||||
if not vars.quiet:
|
if not vars.quiet:
|
||||||
print("{0}{1}{2}".format(colors.CYAN, genout, colors.END))
|
split_gen = genout.split('\n')
|
||||||
|
for line in split_gen:
|
||||||
|
logger.generation(line)
|
||||||
|
|
||||||
# Format output before continuing
|
# Format output before continuing
|
||||||
if not ignore_formatting:
|
if not ignore_formatting:
|
||||||
|
@ -4918,7 +4925,10 @@ def genselect(genout):
|
||||||
# Apply output formatting rules to sequences
|
# Apply output formatting rules to sequences
|
||||||
result["generated_text"] = applyoutputformatting(result["generated_text"])
|
result["generated_text"] = applyoutputformatting(result["generated_text"])
|
||||||
if not vars.quiet:
|
if not vars.quiet:
|
||||||
print("{0}[Result {1}]\n{2}{3}".format(colors.CYAN, i, result["generated_text"], colors.END))
|
logger.info(f"Generation Result {i}")
|
||||||
|
split_gen = result["generated_text"].split('\n')
|
||||||
|
for line in split_gen:
|
||||||
|
logger.generation(line)
|
||||||
i += 1
|
i += 1
|
||||||
|
|
||||||
# Add the options to the actions metadata
|
# Add the options to the actions metadata
|
||||||
|
@ -5257,7 +5267,10 @@ def tpumtjgenerate(txt, minimum, maximum, found_entries=None):
|
||||||
found_entries = tuple(found_entries.copy() for _ in range(vars.numseqs))
|
found_entries = tuple(found_entries.copy() for _ in range(vars.numseqs))
|
||||||
|
|
||||||
if not vars.quiet:
|
if not vars.quiet:
|
||||||
print("{0}Min:{1}, Max:{2}, Txt:{3}{4}".format(colors.YELLOW, minimum, maximum, utils.decodenewlines(tokenizer.decode(txt)), colors.END))
|
logger.debug(f"Prompt Min:{minimum}, Max:{maximum}")
|
||||||
|
split_prompt = utils.decodenewlines(tokenizer.decode(txt)).split('\n')
|
||||||
|
for line in split_prompt:
|
||||||
|
logger.prompt(line)
|
||||||
|
|
||||||
vars._actions = vars.actions
|
vars._actions = vars.actions
|
||||||
vars._prompt = vars.prompt
|
vars._prompt = vars.prompt
|
||||||
|
@ -5345,9 +5358,8 @@ def tpumtjgenerate(txt, minimum, maximum, found_entries=None):
|
||||||
vars.lua_running = False
|
vars.lua_running = False
|
||||||
emit('from_server', {'cmd': 'errmsg', 'data': 'Lua script error; please check console.'}, broadcast=True)
|
emit('from_server', {'cmd': 'errmsg', 'data': 'Lua script error; please check console.'}, broadcast=True)
|
||||||
sendUSStatItems()
|
sendUSStatItems()
|
||||||
print("{0}{1}{2}".format(colors.RED, "***LUA ERROR***: ", colors.END), end="", file=sys.stderr)
|
logger.debug('LUA ERROR: ' + str(e).replace("\033", ""))
|
||||||
print("{0}{1}{2}".format(colors.RED, str(e).replace("\033", ""), colors.END), file=sys.stderr)
|
logger.warning("Lua engine stopped; please open 'Userscripts' and press Load to reinitialize scripts.")
|
||||||
print("{0}{1}{2}".format(colors.YELLOW, "Lua engine stopped; please open 'Userscripts' and press Load to reinitialize scripts.", colors.END), file=sys.stderr)
|
|
||||||
else:
|
else:
|
||||||
emit('from_server', {'cmd': 'errmsg', 'data': 'Error occurred during generator call; please check console.'}, broadcast=True)
|
emit('from_server', {'cmd': 'errmsg', 'data': 'Error occurred during generator call; please check console.'}, broadcast=True)
|
||||||
print("{0}{1}{2}".format(colors.RED, traceback.format_exc().replace("\033", ""), colors.END), file=sys.stderr)
|
print("{0}{1}{2}".format(colors.RED, traceback.format_exc().replace("\033", ""), colors.END), file=sys.stderr)
|
||||||
|
|
Loading…
Reference in New Issue