17d07b280a
Correct gpu_layers
to gpu_blocks
2021-11-14 21:08:49 -05:00
805cb0c8b9
Make sure device_config() still works with all layers on CPU
2021-11-14 18:46:00 -05:00
80aee07816
Use old GPU-only generation if all layers are on the same GPU
...
Apparently, this mode uses less RAM than breakmodel does.
2021-11-14 18:42:18 -05:00
b0ab30cec4
Re-enable GPU-only generation option
2021-11-14 18:24:51 -05:00
3e38b462c6
Hidden Size fix for GPT2 Custom
...
Replaced the JS Hidden Size load with the newer function to fix these models
2021-11-14 16:40:04 +01:00
ecea169553
Improved Unix Support
...
Changes the line-endings to the Unix format and sets KoboldAI to launch with Python3 if executed directly.
(cherry picked from commit 5b0977ceb6807c0f80ce6717891ef5e23c8eeb77)
2021-11-13 21:54:32 -05:00
1596a238f7
Breakmodel automation
...
The only changes are a small addition to the breakmodel section where GPU0 is automatically chosen if the CLI options are used without specifying breakmodel. Lineendings have been changed to Linux formatting for compatibility reasons.
2021-11-14 03:13:52 +01:00
8a916116e3
Remove device=0 because of incompatibility
...
Device=0 breaks some of the pytorch implementations, removed to restore hardware compatibility to 0.16 levels.
2021-11-14 02:33:27 +01:00
4bcffc614e
Allow directly running KoboldAI from CLI in Linux
...
Its made for Python3, so we assume python3 is installed in its usual location. If it isn't you can always run it yourself with whatever command you used prior to this change.
2021-11-14 01:57:43 +01:00
21ae45e9ab
Merge branch 'KoboldAI:main' into united
2021-11-11 17:05:39 +01:00
1fadcbe1e3
Send allowsp command on connect instead of on startup
2021-11-11 00:18:46 -05:00
2fe815e092
Don't broadcast emit calls inside do_connect()
...
This prevents the "thinking" animation from appearing on top of the
submit button under certain circumstances:
* When someone connects to the KoboldAI server while the model is
generating (occurs after generation finishes)
* Occasionally, the browser may suddenly disconnect and reconnect from
Flask-SocketIO during generation, which causes the same problem
2021-11-11 00:14:12 -05:00
11b0291bc4
Use model.transformer.embed_dim if model.transformer.hidden_size doesn't exist
2021-11-10 17:47:14 -05:00
752e19a2bb
Fix vars.modeldim not always being set
2021-11-10 17:38:30 -05:00
2679df9664
Merge branch 'main' into united
2021-11-09 21:33:14 -05:00
c2371cf801
Merge pull request #23 from VE-FORBRYDERNE/scan-test
...
Dynamic world info scan
2021-11-10 03:31:42 +01:00
4af0d9dabd
Merge pull request #78 from VE-FORBRYDERNE/patch
...
Allow remote mode to load from client-side story files
2021-11-06 16:58:05 +01:00
02a56945de
Version bump
2021-11-06 11:50:56 -04:00
bc0f9c8032
Allow remote mode to load from client-side story files
...
(cherry picked from commit a1345263df
)
2021-11-06 11:48:20 -04:00
7c099fe93c
Allow remote mode to load from client-side story files
2021-11-04 19:33:17 -04:00
81bd058caf
Make sure calcsubmitbudget uses the correct reference to vars.actions
2021-11-03 18:57:02 -04:00
a2d7735a51
Dynamic WI scanner should ignore triggers that are already in context
2021-11-03 18:55:53 -04:00
ecfbbdb4a9
Merge branch 'united' into scan-test
2021-11-03 18:23:22 -04:00
0fa47b1249
Fix budget calculation for stories with at least one non-prompt chunk
2021-11-03 18:22:31 -04:00
c11dab894e
Put placeholder variables into calcsubmitbudget
2021-11-03 18:02:19 -04:00
9b18068999
Shallow copy story chunks when generating
2021-11-03 17:53:38 -04:00
b8c3d8c12e
Fix generator output having the wrong length
2021-11-03 16:10:12 -04:00
5b3ce4510f
Make sure that soft_tokens is on the correct device
2021-11-03 16:07:50 -04:00
90fd5a538a
Merge branch 'united' into scan-test
2021-11-03 12:42:18 -04:00
fe2987d894
Fix missing break statement in device_config
2021-11-03 12:42:04 -04:00
bd76ab333c
Set numseqs to 1 if using dynamic world info scan
2021-11-03 12:28:17 -04:00
0a91ea27b3
Make the dynamic world info scan toggleable
2021-11-03 12:18:48 -04:00
de3664e73c
Add an assertion for the value of already_generated
2021-11-03 12:01:45 -04:00
ec8ec55256
Dynamic world info scan
2021-11-03 11:54:48 -04:00
aa998ba5e9
Merge pull request #20 from VE-FORBRYDERNE/sp
...
Soft prompt support for PyTorch models
2021-10-30 00:35:44 +02:00
206c01008e
Fix budget calculation when using soft prompt
2021-10-29 11:44:51 -04:00
c9c370aa17
Merge branch 'KoboldAI:main' into united
2021-10-28 23:29:29 +02:00
bf4e7742ac
Patch GPTJForCausalLM, if it exists, to support soft prompting
2021-10-28 17:18:28 -04:00
40b4631f6c
Clamp input_ids in place
...
Apparently transformers maintains an internal reference to input_ids
(to use for repetition penalty) so we have to clamp the internal
version, too, because otherwise transformers will throw an out-of-bounds
error upon attempting to access token IDs that are not in the
vocabulary.
2021-10-28 16:52:39 -04:00
24d5d63c9f
Use the correct generation min and max when using soft prompt
2021-10-28 16:39:59 -04:00
511817132a
Don't change the shape of transformer.wte
2021-10-28 15:39:59 -04:00
a1ae11630a
Make sure to cast vars.sp to the correct dtype
2021-10-28 13:22:07 -04:00
1556bd32a5
Use torch.where to inject the soft prompt instead of torch.cat
2021-10-28 13:20:14 -04:00
248e0bd24b
Fix soft prompt loading code
2021-10-28 00:29:42 -04:00
4e3cc93020
Merge branch 'united' into sp
2021-10-23 11:45:03 -04:00
7b73d7cfdd
Single Line Mode
...
Adds Single Line mode, optimized for things like chatbot testing and other cases where you want to have control over what happens after a paragraph.
This can also be used as a foundation for a chatbot optimized interface mode.
2021-10-23 17:30:48 +02:00
1f449a9dda
Soft prompt support (6B Colabs not supported yet)
2021-10-22 14:18:10 -04:00
3501f03153
Create settings directory if it doesn't exist when using InferKit/OAI
2021-10-21 23:33:32 -04:00
fa0f8af1d6
Merge branch 'KoboldAI:main' into united
2021-10-15 08:23:06 +02:00
9513240dfb
Version bump
...
Since VE fixed important things in the editor i want users to be able to see this easier
2021-10-15 08:22:32 +02:00