mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
Fix llama loading
This commit is contained in:
10
aiserver.py
10
aiserver.py
@@ -2915,10 +2915,10 @@ def load_model(use_gpu=True, gpu_layers=None, disk_layers=None, initial_load=Fal
|
|||||||
|
|
||||||
@functools.lru_cache(maxsize=None)
|
@functools.lru_cache(maxsize=None)
|
||||||
def get_original_key(key):
|
def get_original_key(key):
|
||||||
# try:
|
try:
|
||||||
return max((original_key for original_key in utils.module_names if original_key.endswith(key)), key=len)
|
return max((original_key for original_key in utils.module_names if original_key.endswith(key)), key=len)
|
||||||
# except ValueError:
|
except ValueError:
|
||||||
# return key
|
return key
|
||||||
|
|
||||||
for key, value in model_dict.items():
|
for key, value in model_dict.items():
|
||||||
original_key = get_original_key(key)
|
original_key = get_original_key(key)
|
||||||
@@ -3128,8 +3128,8 @@ def load_model(use_gpu=True, gpu_layers=None, disk_layers=None, initial_load=Fal
|
|||||||
path_4bit = os.path.join(koboldai_vars.custmodpth, "4bit.pt")
|
path_4bit = os.path.join(koboldai_vars.custmodpth, "4bit.pt")
|
||||||
|
|
||||||
if not os.path.isfile(path_4bit):
|
if not os.path.isfile(path_4bit):
|
||||||
print(f"4-bit file {path_4bit} not found, aborting 4-bit load")
|
print(f"4-bit file {path_4bit} not found, loading failed")
|
||||||
use_4_bit = False
|
raise RuntimeError(f"4-bit load failed. PT-File not found at {path_4bit}")
|
||||||
|
|
||||||
if use_4_bit:
|
if use_4_bit:
|
||||||
print(f"Trying to load {koboldai_vars.model_type} model in 4-bit")
|
print(f"Trying to load {koboldai_vars.model_type} model in 4-bit")
|
||||||
|
Reference in New Issue
Block a user