diff --git a/aiserver.py b/aiserver.py index fc8698f0..cb274537 100644 --- a/aiserver.py +++ b/aiserver.py @@ -1405,6 +1405,7 @@ def load_model(use_gpu=True, gpu_layers=None, initial_load=False, online_model=" global generator global torch global model_config + global GPT2TokenizerFast print("Loading vars.model: {} vars.custmodpth: {}".format(vars.model, vars.custmodpth)) vars.noai = False if not initial_load: