mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-03-10 00:20:11 +01:00
Fix for loading models that don't support breakmodel (GPU/CPU support in UI)
This commit is contained in:
parent
168ae8083c
commit
be719a7e5e
13
aiserver.py
13
aiserver.py
@ -2230,7 +2230,7 @@ def load_model(use_gpu=True, gpu_layers=None, disk_layers=None, initial_load=Fal
|
|||||||
loadsettings()
|
loadsettings()
|
||||||
logger.init("GPU support", status="Searching")
|
logger.init("GPU support", status="Searching")
|
||||||
vars.hascuda = torch.cuda.is_available()
|
vars.hascuda = torch.cuda.is_available()
|
||||||
vars.bmsupported = (utils.HAS_ACCELERATE or vars.model_type in ("gpt_neo", "gptj", "xglm", "opt")) and not vars.nobreakmodel
|
vars.bmsupported = ((utils.HAS_ACCELERATE and vars.model_type != 'gpt2') or vars.model_type in ("gpt_neo", "gptj", "xglm", "opt")) and not vars.nobreakmodel
|
||||||
if(args.breakmodel is not None and args.breakmodel):
|
if(args.breakmodel is not None and args.breakmodel):
|
||||||
logger.warning("--breakmodel is no longer supported. Breakmodel mode is now automatically enabled when --breakmodel_gpulayers is used (see --help for details).")
|
logger.warning("--breakmodel is no longer supported. Breakmodel mode is now automatically enabled when --breakmodel_gpulayers is used (see --help for details).")
|
||||||
if(args.breakmodel_layers is not None):
|
if(args.breakmodel_layers is not None):
|
||||||
@ -2256,7 +2256,7 @@ def load_model(use_gpu=True, gpu_layers=None, disk_layers=None, initial_load=Fal
|
|||||||
vars.breakmodel = True
|
vars.breakmodel = True
|
||||||
else:
|
else:
|
||||||
vars.breakmodel = False
|
vars.breakmodel = False
|
||||||
vars.usegpu = True
|
vars.usegpu = use_gpu
|
||||||
|
|
||||||
|
|
||||||
# Ask for API key if InferKit was selected
|
# Ask for API key if InferKit was selected
|
||||||
@ -2430,9 +2430,14 @@ def load_model(use_gpu=True, gpu_layers=None, disk_layers=None, initial_load=Fal
|
|||||||
yield False
|
yield False
|
||||||
|
|
||||||
# If custom GPT2 model was chosen
|
# If custom GPT2 model was chosen
|
||||||
if(vars.model == "GPT2Custom"):
|
if(vars.model_type == "gpt2"):
|
||||||
vars.lazy_load = False
|
vars.lazy_load = False
|
||||||
model_config = open(vars.custmodpth + "/config.json", "r")
|
if os.path.exists(vars.custmodpth):
|
||||||
|
model_config = open(vars.custmodpth + "/config.json", "r")
|
||||||
|
elif os.path.exists(os.path.join("models/", vars.custmodpth)):
|
||||||
|
config_path = os.path.join("models/", vars.custmodpth)
|
||||||
|
config_path = os.path.join(config_path, "config.json").replace("\\", "//")
|
||||||
|
model_config = open(config_path, "r")
|
||||||
js = json.load(model_config)
|
js = json.load(model_config)
|
||||||
with(maybe_use_float16()):
|
with(maybe_use_float16()):
|
||||||
try:
|
try:
|
||||||
|
Loading…
x
Reference in New Issue
Block a user