Commit Graph

4221 Commits

Author SHA1 Message Date
ebolam
7e0778c871 Remove extra debug stuff 2023-05-19 09:14:37 -04:00
ebolam
36e679b366 Merge branch 'Model_Plugins' of https://github.com/ebolam/KoboldAI into Model_Plugins 2023-05-19 09:11:22 -04:00
ebolam
99cffd4755 Colab GPU edition fixes 2023-05-19 09:11:08 -04:00
ebolam
3db231562f Merge pull request #382 from henk717/united
Update to united
2023-05-19 06:05:25 -04:00
ebolam
56d2705f4b removed breakmodel command line arguments (except nobreakmodel) 2023-05-18 20:19:33 -04:00
ebolam
06f59a7b7b Moved model backends to separate folders
added some model backend settings save/load
2023-05-18 20:14:33 -04:00
ebolam
4040538d34 Model Backends now defined in the menu 2023-05-18 18:34:00 -04:00
ebolam
182ecff202 Added in model backend to the command line arguments 2023-05-18 16:01:17 -04:00
0cc4m
2c18d9f2b5 Update GPTQ module to 0.0.3 2023-05-18 21:51:03 +02:00
ebolam
f027d8b6e5 Better working valid detection and named model backends for UI 2023-05-17 21:15:31 -04:00
Henk
b2501e4693 4.29 was still to buggy 2023-05-16 22:15:59 +02:00
Henk
59c96b5b7a Unban fix 2023-05-15 22:38:12 +02:00
Henk
c5100b4eab Unban Tensor 2023-05-15 22:21:22 +02:00
Henk
56443bc7ea Unban torch._tensor._rebuild_tensor_v2 2023-05-15 21:44:01 +02:00
0cc4m
3d4d5df76b Remove rocm wheel, because it didn't work correctly 2023-05-13 20:33:13 +02:00
0cc4m
7f7b350741 Catch further error during multigpu 4bit setup 2023-05-13 20:31:01 +02:00
Henk
205c64f1ea More universal pytorch folder detection 2023-05-13 20:26:55 +02:00
0cc4m
266c0574f6 Fix 4bit pt loading, add traceback output to GPT2 fallback 2023-05-13 20:15:11 +02:00
ebolam
c6b17889d0 Updated to latest united 2023-05-12 07:53:27 -04:00
Henk
67df9b917f Reintroduce 4.29 Transformers 2023-05-12 09:08:07 +02:00
henk717
db32aba74d Merge pull request #359 from one-some/gptj-fix
GPT-J fix
2023-05-12 08:40:00 +02:00
ebolam
aaa9133899 Disk Cache working
UI valid marker broken for disk cache
2023-05-11 21:22:33 -04:00
ebolam
a6f0e97ba0 Working(?) breakmodel 2023-05-11 20:40:05 -04:00
ebolam
69d942c00c Kind of working breakmodel 2023-05-11 20:22:30 -04:00
somebody
3065c1b40e Ignore missing keys in get_original_key 2023-05-11 17:10:43 -05:00
somebody
c16336f646 Add traceback to debug log on fallback 2023-05-11 17:10:19 -05:00
somebody
6838563ea9 Merge branch 'united' of https://github.com/henk717/KoboldAI into united 2023-05-11 16:32:25 -05:00
ebolam
a9c785d0f0 Fix for Horde 2023-05-11 14:20:14 -04:00
ebolam
e9c845dc2a Fix for badwordIDs 2023-05-11 14:14:52 -04:00
Henk
20b54eb9ff Revert 4.29 due to unforseen consequences 2023-05-11 19:06:39 +02:00
ebolam
4605d10c37 Next iteration. Model Loading is broken completely now :) 2023-05-11 12:08:35 -04:00
Henk
edd9c7d782 Warning polish 2023-05-11 15:13:59 +02:00
ebolam
77dd5aa725 Minor update 2023-05-11 09:09:09 -04:00
Henk
e932364a1e RWKV support 2023-05-11 14:56:12 +02:00
Henk
84e4cb0f4a Update Transformers 2023-05-11 13:44:53 +02:00
somebody
546ba84723 Fix memory->genre bug in context viewer bar tooltip
Crazy change I know
2023-05-10 19:10:23 -05:00
ebolam
71aee4dbd8 First concept of model plugins with a conceptual UI.
Completely breaks UI2 model loading.
2023-05-10 16:30:46 -04:00
0cc4m
a2d01bb9e4 Update to GPTQ module 0.0.2, add support for upstream cuda quantizations, automatic detection 2023-05-09 22:20:35 +02:00
Henk
702f59b2db Downgrade ROCM properly 2023-05-09 22:10:01 +02:00
Henk
9fdc2f73a6 ROCM Downgrade for stability 2023-05-09 20:59:10 +02:00
henk717
a649bd8d3c Merge pull request #357 from one-some/tpu-api-fix
Fix TPU API errors
2023-05-09 01:05:20 +02:00
somebody
a9e342ca64 Fix TPU API errors 2023-05-08 17:34:59 -05:00
0cc4m
6121598142 Fix multigpu loading without lazy-loader 2023-05-08 22:57:09 +02:00
0cc4m
4f94247910 Fix chat mode empty generation error 2023-05-08 22:56:17 +02:00
0cc4m
e55a9d31c2 Update readme, clean up gitmodules file 2023-05-08 22:55:59 +02:00
henk717
0f9129859f Merge pull request #353 from Zurnaz/llama_tpu_tokenizer_fix
fix: tpu tokenizers errors
2023-05-08 19:41:33 +02:00
Bogdan Drema
d53726bed6 fix: tpu tokenizers errors 2023-05-08 18:24:34 +01:00
henk717
cb4af7e56e Update requirements_mtj.txt 2023-05-08 17:23:49 +02:00
0cc4m
6b4d3218d6 Fix OOM when loading large model split across GPUs 2023-05-07 06:55:51 +02:00
0cc4m
51e6dcdcd4 Revert accidental install_requirements change 2023-05-07 06:42:32 +02:00