mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
Fix horde tokenizer
This commit is contained in:
@@ -30,6 +30,7 @@ class model_backend(InferenceModel):
|
|||||||
self.key = "0000000000"
|
self.key = "0000000000"
|
||||||
self.models = self.get_cluster_models()
|
self.models = self.get_cluster_models()
|
||||||
self.model_name = "Horde"
|
self.model_name = "Horde"
|
||||||
|
self.model = []
|
||||||
|
|
||||||
|
|
||||||
# Do not allow API to be served over the API
|
# Do not allow API to be served over the API
|
||||||
@@ -114,7 +115,7 @@ class model_backend(InferenceModel):
|
|||||||
|
|
||||||
engines = req.json()
|
engines = req.json()
|
||||||
try:
|
try:
|
||||||
engines = [{"text": "all", "value": "all"}] + [{"text": en["name"], "value": en["name"]} for en in engines]
|
engines = [{"text": "All", "value": "all"}] + [{"text": en["name"], "value": en["name"]} for en in engines]
|
||||||
except:
|
except:
|
||||||
logger.error(engines)
|
logger.error(engines)
|
||||||
raise
|
raise
|
||||||
@@ -127,10 +128,14 @@ class model_backend(InferenceModel):
|
|||||||
return engines
|
return engines
|
||||||
|
|
||||||
def _load(self, save_model: bool, initial_load: bool) -> None:
|
def _load(self, save_model: bool, initial_load: bool) -> None:
|
||||||
|
tokenizer_name = "gpt2"
|
||||||
|
if len(self.model) > 0:
|
||||||
|
if self.model[0] == "all" and len(self.model) > 1:
|
||||||
|
tokenizer_name = self.model[1]
|
||||||
|
else:
|
||||||
|
tokenizer_name = self.model[0]
|
||||||
self.tokenizer = self._get_tokenizer(
|
self.tokenizer = self._get_tokenizer(
|
||||||
self.model
|
tokenizer_name
|
||||||
#if len(self.model) > 0
|
|
||||||
#else "gpt2",
|
|
||||||
)
|
)
|
||||||
|
|
||||||
def _save_settings(self):
|
def _save_settings(self):
|
||||||
|
Reference in New Issue
Block a user