Fix for bug that crashes generation when using soft prompts

This commit is contained in:
ebolam
2023-01-09 19:56:49 -05:00
parent 3b551ec7ab
commit 4668bc8ec0

View File

@@ -515,6 +515,7 @@ class koboldai_vars(object):
else: else:
tokens = [] tokens = []
for item in context: for item in context:
if item['type'] != 'soft_prompt':
tokens.extend([x[0] for x in item['tokens']]) tokens.extend([x[0] for x in item['tokens']])
if send_context: if send_context:
@@ -524,6 +525,9 @@ class koboldai_vars(object):
logger.debug("Token Budget: {}. Used Tokens: {}".format(token_budget, used_tokens)) logger.debug("Token Budget: {}. Used Tokens: {}".format(token_budget, used_tokens))
if return_text: if return_text:
return "".join([x['text'] for x in context]) return "".join([x['text'] for x in context])
print(self.sp_length)
print(context)
print(tokens)
return tokens, used_tokens, used_tokens+self.genamt, set(used_world_info) return tokens, used_tokens, used_tokens+self.genamt, set(used_world_info)
def is_model_torch(self) -> bool: def is_model_torch(self) -> bool: