mirror of
https://github.com/KoboldAI/KoboldAI-Client.git
synced 2025-04-15 02:42:15 +02:00
Reduced default max_length parameter to 512.
Added warning about VRAM usage to Max Tokens tooltip.
This commit is contained in:
parent
d632976fbf
commit
ba1ba0fc8a
@ -48,7 +48,7 @@ class vars:
|
|||||||
model = ""
|
model = ""
|
||||||
noai = False # Runs the script without starting up the transformers pipeline
|
noai = False # Runs the script without starting up the transformers pipeline
|
||||||
aibusy = False # Stops submissions while the AI is working
|
aibusy = False # Stops submissions while the AI is working
|
||||||
max_length = 1024 # Maximum number of tokens to submit per action
|
max_length = 512 # Maximum number of tokens to submit per action
|
||||||
ikmax = 3000 # Maximum number of characters to submit to InferKit
|
ikmax = 3000 # Maximum number of characters to submit to InferKit
|
||||||
genamt = 60 # Amount of text for each action to generate
|
genamt = 60 # Amount of text for each action to generate
|
||||||
ikgen = 200 # Number of characters for InferKit to generate
|
ikgen = 200 # Number of characters for InferKit to generate
|
||||||
|
@ -29,15 +29,15 @@ gensettingstf = [{
|
|||||||
"max": 2.0,
|
"max": 2.0,
|
||||||
"step": 0.05,
|
"step": 0.05,
|
||||||
"default": 1.0,
|
"default": 1.0,
|
||||||
"tooltip": "Used to penalize words that were already generated or belong to the context"
|
"tooltip": "Used to penalize words that were already generated or belong to the context."
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"uitype": "slider",
|
"uitype": "slider",
|
||||||
"unit": "int",
|
"unit": "int",
|
||||||
"label": "Amount to Generate",
|
"label": "Amount to Generate",
|
||||||
"id": "setoutput",
|
"id": "setoutput",
|
||||||
"min": 10,
|
"min": 16,
|
||||||
"max": 500,
|
"max": 512,
|
||||||
"step": 2,
|
"step": 2,
|
||||||
"default": 60,
|
"default": 60,
|
||||||
"tooltip": "Number of tokens the AI should generate. Higher numbers will take longer to generate."
|
"tooltip": "Number of tokens the AI should generate. Higher numbers will take longer to generate."
|
||||||
@ -50,8 +50,8 @@ gensettingstf = [{
|
|||||||
"min": 512,
|
"min": 512,
|
||||||
"max": 2048,
|
"max": 2048,
|
||||||
"step": 8,
|
"step": 8,
|
||||||
"default": 1024,
|
"default": 512,
|
||||||
"tooltip": "Number of tokens of context to submit to the AI for sampling."
|
"tooltip": "Max number of tokens of context to submit to the AI for sampling. Make sure this is higher than Amount to Generate. Higher values increase VRAM/RAM usage."
|
||||||
}]
|
}]
|
||||||
|
|
||||||
gensettingsik =[{
|
gensettingsik =[{
|
||||||
|
Loading…
x
Reference in New Issue
Block a user