From 2455e5d60cbf3a0d40b72bf36d27f0cb6e978b1f Mon Sep 17 00:00:00 2001 From: Henk Date: Mon, 11 Sep 2023 01:21:29 +0200 Subject: [PATCH] Hide exllama if not installed --- modeling/inference_models/exllama/class.py | 61 ++++++++++++---------- 1 file changed, 33 insertions(+), 28 deletions(-) diff --git a/modeling/inference_models/exllama/class.py b/modeling/inference_models/exllama/class.py index 33d2fcc2..4539b7a3 100644 --- a/modeling/inference_models/exllama/class.py +++ b/modeling/inference_models/exllama/class.py @@ -1,36 +1,40 @@ from __future__ import annotations +try: + import time, json + import torch + import requests + import numpy as np + from typing import List, Optional, Union + import os + import glob + from pathlib import Path + import re + import warnings + import gc -import time, json -import torch -import requests -import numpy as np -from typing import List, Optional, Union -import os -import glob -from pathlib import Path -import re -import warnings -import gc + import utils + from logger import logger -import utils -from logger import logger + from modeling import warpers + from modeling.warpers import Warper + from modeling.stoppers import Stoppers + from modeling.post_token_hooks import PostTokenHooks + from modeling.inference_model import ( + GenerationResult, + GenerationSettings, + InferenceModel, + ModelCapabilities, + ) -from modeling import warpers -from modeling.warpers import Warper -from modeling.stoppers import Stoppers -from modeling.post_token_hooks import PostTokenHooks -from modeling.inference_model import ( - GenerationResult, - GenerationSettings, - InferenceModel, - ModelCapabilities, -) + from modeling.tokenizer import GenericTokenizer -from modeling.tokenizer import GenericTokenizer -from exllama.model import ExLlama, ExLlamaCache, ExLlamaConfig -from transformers import LlamaTokenizer -from exllama.generator import ExLlamaGenerator + from exllama.model import ExLlama, ExLlamaCache, ExLlamaConfig + from transformers import LlamaTokenizer + from exllama.generator import ExLlamaGenerator + load_failed = False +except: + load_failed = True model_backend_type = "GPTQ" model_backend_name = "ExLlama" @@ -101,10 +105,11 @@ class model_backend(InferenceModel): stopper_hooks=True, post_token_probs=False, ) + self.disable = load_failed def is_valid(self, model_name, model_path, menu_path): - gptq_model, _ = load_model_gptq_settings(model_path) try: + gptq_model, _ = load_model_gptq_settings(model_path) self.model_config = self._load_config(model_name, model_path) return self.model_config and gptq_model except: