mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
Fix for exllama (v1 and v2) showing 2x status (0-200%) on generation
This commit is contained in:
@@ -340,7 +340,8 @@ class model_backend(InferenceModel):
|
|||||||
|
|
||||||
self._post_token_gen(self.generator.sequence)
|
self._post_token_gen(self.generator.sequence)
|
||||||
|
|
||||||
utils.koboldai_vars.generated_tkns += 1
|
#This is taken care of in the core stopper class that's called below. If you're not using core stoppers then it should remain here
|
||||||
|
#utils.koboldai_vars.generated_tkns += 1
|
||||||
|
|
||||||
# Apply stoppers
|
# Apply stoppers
|
||||||
do_stop = False
|
do_stop = False
|
||||||
|
@@ -315,7 +315,8 @@ class model_backend(InferenceModel):
|
|||||||
|
|
||||||
self._post_token_gen(self.generator.sequence_ids)
|
self._post_token_gen(self.generator.sequence_ids)
|
||||||
|
|
||||||
utils.koboldai_vars.generated_tkns += 1
|
#This is taken care of in the core stopper class that's called below. If you're not using core stoppers then it should remain here
|
||||||
|
#utils.koboldai_vars.generated_tkns += 1
|
||||||
|
|
||||||
# Apply stoppers
|
# Apply stoppers
|
||||||
do_stop = False
|
do_stop = False
|
||||||
|
Reference in New Issue
Block a user