mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-06-05 21:59:24 +02:00
Fix for exllama (v1 and v2) showing 2x status (0-200%) on generation
This commit is contained in:
@@ -340,7 +340,8 @@ class model_backend(InferenceModel):
|
||||
|
||||
self._post_token_gen(self.generator.sequence)
|
||||
|
||||
utils.koboldai_vars.generated_tkns += 1
|
||||
#This is taken care of in the core stopper class that's called below. If you're not using core stoppers then it should remain here
|
||||
#utils.koboldai_vars.generated_tkns += 1
|
||||
|
||||
# Apply stoppers
|
||||
do_stop = False
|
||||
|
@@ -315,7 +315,8 @@ class model_backend(InferenceModel):
|
||||
|
||||
self._post_token_gen(self.generator.sequence_ids)
|
||||
|
||||
utils.koboldai_vars.generated_tkns += 1
|
||||
#This is taken care of in the core stopper class that's called below. If you're not using core stoppers then it should remain here
|
||||
#utils.koboldai_vars.generated_tkns += 1
|
||||
|
||||
# Apply stoppers
|
||||
do_stop = False
|
||||
|
Reference in New Issue
Block a user