Commit Graph

4188 Commits

Author SHA1 Message Date
somebody
38c53191d3 possible fix for cache dl thing 2023-04-14 20:25:03 -05:00
ebolam
be0319cf03 added VERSION link in the left flyout menu footer that will show which git REPO and BRANCH the system is running on.
Fix for pasting text into the game area
Fix for ctrl+a then delete
Fix for changing text then back/forward a bunch loosing text
2023-04-14 13:54:11 -04:00
Henk
bde9c6980f Transformers 4.28 support 2023-04-14 14:13:46 +02:00
henk717
03ab4b25af Merge branch 'KoboldAI:main' into united 2023-04-14 13:59:55 +02:00
biscober
35f908e147 Update install_requirements.bat (#7)
* Update install_requirements.bat

move command to dismount temp B drive to after pip install command which requires B drive to still be mounted

* Update install_requirements.bat

cmd /k not necessary

* Update install_requirements.bat

add quotes (probably not required but w/e)
2023-04-11 04:37:48 +02:00
0cc4m
687d107d20 Update README, remove steps that are no longer required 2023-04-10 22:46:12 +02:00
0cc4m
b628aec719 Automatic installation of the quant_cuda module during install_requirements
Kepler (K40+) and Maxwell support
2023-04-10 22:37:16 +02:00
henk717
2385a34098 Merge pull request #325 from YellowRoseCx/patch-1
Add IP Whitelisting to --host
2023-04-10 14:08:09 +02:00
somebody
334c09606b Fix for tokenizer stuff on pythia 2023-04-09 18:23:58 -05:00
somebody
3e8e3a18b0 Fix for custom gpt2 2023-04-09 18:23:52 -05:00
somebody
f73a8bb808 Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv 2023-04-09 14:38:09 -05:00
somebody
fedbffd07b Small fixes
Typos galore!
2023-04-09 13:35:28 -05:00
0cc4m
7efd314428 Improve guide 2023-04-07 20:10:24 +02:00
0cc4m
ccf34a5edc Fix merge issues with upstream, merge changes 2023-04-07 19:51:07 +02:00
0cc4m
636c4e5a52 Update gptq repo 2023-04-07 11:48:57 +02:00
YellowRoseCx
ac98cd6dd1 add IP_whitelisting to koboldai_settings.py 2023-04-05 21:27:59 -05:00
YellowRoseCx
71e5d23a5b Add IP whitelisting to --host 2023-04-05 21:23:24 -05:00
0cc4m
40092cc9fa Improve guide formatting 2023-04-05 21:49:13 +02:00
0cc4m
8b4375307c Update file formatting section in guide 2023-04-05 21:10:40 +02:00
0cc4m
e4f8a9344c Merge pull request #1 from Digitous/patch-1
Add install instructions
2023-04-05 21:08:14 +02:00
Henk
80e4b9e536 Merge branch 'main' into united 2023-04-05 00:22:30 +02:00
henk717
29c2d4b7a6 Removing Pygmalion from the TPU colab to get it unbanned 2023-04-04 19:51:18 +02:00
henk717
fd12214091 Clean the description of the GPU colab 2023-04-04 19:40:22 +02:00
henk717
bb51127bbf We no longer support Pygmalion on Colab due to Google's Pygmalion ban 2023-04-04 19:37:15 +02:00
Henk
4b71da1714 Horde settings in the UI 2023-04-04 17:20:43 +02:00
0cc4m
ce6761e744 Fix issue causing expected scalar type Float but found Half RuntimeErrors 2023-04-04 07:46:53 +02:00
Henk
8bf533da9a Pin Accelerate Version 2023-04-04 01:47:59 +02:00
somebody
8412f83ce5 Breakmodel: Fix typo 2023-04-03 18:41:18 -05:00
0cc4m
b9df9b6f59 Improve CPU offloading speed significantly when offloading less than half of the layers 2023-04-03 20:27:17 +02:00
0cc4m
5abdecad2c Merge pull request #5 from 0cc4m/cpu-offload-1
CPU Offloading Support
2023-04-03 06:52:48 +02:00
0cc4m
ec4177a6d6 Remove cudatoolkit-dev and gcc/gxx 9 from conda env because they didn't resolve on Windows 2023-04-03 06:50:36 +02:00
somebody
4230fe4229 Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv 2023-04-02 16:41:21 -05:00
somebody
77f0797b1a Model fix 2023-04-02 15:47:52 -05:00
somebody
9d70646e4d Lazyload: Safetensors 2023-04-02 15:40:34 -05:00
0cc4m
c8d00b7a10 Add CPU offloading support for GPT-NeoX, GPT-J and OPT 2023-04-02 18:36:31 +02:00
0cc4m
e742083703 Fix multi-gpu-offloading 2023-04-02 11:17:29 +02:00
0cc4m
2729b77640 Add offload.py adapted from llama_inference_offload.py, with multi-gpu support and some improvements. Not yet functional, and still just supports Llama 2023-04-02 10:32:19 +02:00
Henk
4a8b099888 Model loading fix 2023-04-02 00:29:56 +02:00
0cc4m
110f8229c5 Add cudatoolkit-dev for compilation, compatible gcc 9 and update transformers to fix error in gptq 2023-04-01 21:33:05 +02:00
0cc4m
bf0c999412 Update GPTQ to support AMD 2023-04-01 14:19:51 +02:00
0cc4m
d3a5ca6505 Update gptq submodule to latest 2023-04-01 08:52:08 +00:00
0cc4m
6eae457479 Fix 4bit groupsize param letter
Use g instead of b for groupsize name, for example 4bit-128g.safetensors
2023-03-31 15:36:03 +02:00
henk717
72b4669563 Fix the chex dependency 2023-03-30 23:41:35 +02:00
0cc4m
aa2292b3a4 Enable multi-gpu support 2023-03-30 19:40:49 +02:00
0cc4m
61b13604b6 Fix bug in 4-bit load fallback 2023-03-30 10:57:04 +02:00
henk717
943d0fe68a Merge branch 'KoboldAI:main' into united 2023-03-30 00:51:17 +02:00
henk717
ab779efe0e Merge pull request #276 from YellowRoseCx/stable-branch
Update README and remove unavailable model from gpu.ipynb
2023-03-30 00:50:15 +02:00
YellowRoseCx
3c48a77a52 Update README.md
changed Colab GPU models listed to their higher quality counter parts
2023-03-29 17:44:44 -05:00
YellowRoseCx
f826930c02 Update GPU.ipynb
removed litv2-6B-rev3
2023-03-29 17:41:01 -05:00
0cc4m
9d0477f5f7 Fix bug where it picks old model despite new one available 2023-03-29 22:05:44 +00:00