Fix llama loading

This commit is contained in:
0cc4m
2023-03-21 21:58:31 +00:00
parent ecd065a881
commit c7edc764b9

View File

@@ -2915,10 +2915,10 @@ def load_model(use_gpu=True, gpu_layers=None, disk_layers=None, initial_load=Fal
@functools.lru_cache(maxsize=None) @functools.lru_cache(maxsize=None)
def get_original_key(key): def get_original_key(key):
# try: try:
return max((original_key for original_key in utils.module_names if original_key.endswith(key)), key=len) return max((original_key for original_key in utils.module_names if original_key.endswith(key)), key=len)
# except ValueError: except ValueError:
# return key return key
for key, value in model_dict.items(): for key, value in model_dict.items():
original_key = get_original_key(key) original_key = get_original_key(key)
@@ -3128,8 +3128,8 @@ def load_model(use_gpu=True, gpu_layers=None, disk_layers=None, initial_load=Fal
path_4bit = os.path.join(koboldai_vars.custmodpth, "4bit.pt") path_4bit = os.path.join(koboldai_vars.custmodpth, "4bit.pt")
if not os.path.isfile(path_4bit): if not os.path.isfile(path_4bit):
print(f"4-bit file {path_4bit} not found, aborting 4-bit load") print(f"4-bit file {path_4bit} not found, loading failed")
use_4_bit = False raise RuntimeError(f"4-bit load failed. PT-File not found at {path_4bit}")
if use_4_bit: if use_4_bit:
print(f"Trying to load {koboldai_vars.model_type} model in 4-bit") print(f"Trying to load {koboldai_vars.model_type} model in 4-bit")