mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
Modeling: Maybe fix samplers
This commit is contained in:
@@ -10,6 +10,7 @@ import utils
|
|||||||
import koboldai_settings
|
import koboldai_settings
|
||||||
from logger import logger, Colors
|
from logger import logger, Colors
|
||||||
|
|
||||||
|
from modeling import warpers
|
||||||
from modeling.inference_model import (
|
from modeling.inference_model import (
|
||||||
GenerationResult,
|
GenerationResult,
|
||||||
GenerationSettings,
|
GenerationSettings,
|
||||||
@@ -262,6 +263,8 @@ class HFMTJInferenceModel(HFInferenceModel):
|
|||||||
seed: Optional[int] = None,
|
seed: Optional[int] = None,
|
||||||
**kwargs,
|
**kwargs,
|
||||||
) -> GenerationResult:
|
) -> GenerationResult:
|
||||||
|
warpers.update_settings()
|
||||||
|
|
||||||
soft_tokens = self.get_soft_tokens()
|
soft_tokens = self.get_soft_tokens()
|
||||||
|
|
||||||
dynamic_inference = kwargs.get("tpu_dynamic_inference", False)
|
dynamic_inference = kwargs.get("tpu_dynamic_inference", False)
|
||||||
|
Reference in New Issue
Block a user