mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
RWKV: Move import right before usage
So we don't needlessly compile the cuda kernel
This commit is contained in:
@@ -592,7 +592,6 @@ from modeling.inference_models.legacy_gpt2_hf import CustomGPT2HFTorchInferenceM
|
|||||||
from modeling.inference_models.hf_mtj import HFMTJInferenceModel
|
from modeling.inference_models.hf_mtj import HFMTJInferenceModel
|
||||||
from modeling.inference_models.horde import HordeInferenceModel
|
from modeling.inference_models.horde import HordeInferenceModel
|
||||||
from modeling.inference_models.openai import OpenAIAPIInferenceModel
|
from modeling.inference_models.openai import OpenAIAPIInferenceModel
|
||||||
from modeling.inference_models.rwkv import RWKVInferenceModel
|
|
||||||
|
|
||||||
|
|
||||||
old_socketio_on = socketio.on
|
old_socketio_on = socketio.on
|
||||||
@@ -1894,6 +1893,7 @@ def load_model(use_gpu=True, gpu_layers=None, disk_layers=None, initial_load=Fal
|
|||||||
elif "rwkv" in koboldai_vars.model:
|
elif "rwkv" in koboldai_vars.model:
|
||||||
if koboldai_vars.use_colab_tpu:
|
if koboldai_vars.use_colab_tpu:
|
||||||
raise RuntimeError("RWKV is not supported on the TPU.")
|
raise RuntimeError("RWKV is not supported on the TPU.")
|
||||||
|
from modeling.inference_models.rwkv import RWKVInferenceModel
|
||||||
model = RWKVInferenceModel(koboldai_vars.model)
|
model = RWKVInferenceModel(koboldai_vars.model)
|
||||||
model.load()
|
model.load()
|
||||||
elif not koboldai_vars.use_colab_tpu and not koboldai_vars.noai:
|
elif not koboldai_vars.use_colab_tpu and not koboldai_vars.noai:
|
||||||
|
Reference in New Issue
Block a user