one-some
1b500c7179
Merge pull request #5 from LostRuins/concedo_api
...
Added stop sequences functionality for API calls
2023-04-15 10:51:31 -05:00
somebody
2b950f08d3
Remove legacy no accelerate fallback code
...
Was causing issues with disk cache the old code had a
`and not utils.HAS_ACCELERATE` preceding it (a variable which no longer
exists), and since disk cache is accelerate only, there was no disk
handling code in here. Anyway its bad so blast it
2023-04-15 10:47:31 -05:00
Henk
67334bd698
Pin accelerate version
2023-04-15 17:45:00 +02:00
somebody
b2e6fcfe3a
Remove line that sets disk_layers to None always
...
whoops
2023-04-15 10:41:10 -05:00
ebolam
348df9ddce
Remove bits and bytes check so it doesn't throw warning on windows
2023-04-15 10:12:20 -04:00
Henk
3eda7269f7
Fix incorrect host merge
2023-04-15 14:58:24 +02:00
Concedo
dd01cf1a93
Merge branch 'concedo_api' of https://github.com/LostRuins/KoboldAI into concedo_api
2023-04-15 18:10:28 +08:00
Concedo
9705b7b79c
increase API version (+1 squashed commits)
...
Squashed commits:
[c168c08
] Added stop sequences functionality for API calls
2023-04-15 18:09:53 +08:00
Concedo
d22423e4be
increase API version
2023-04-15 18:09:29 +08:00
Concedo
c168c08245
Added stop sequences functionality for API calls
2023-04-15 18:00:11 +08:00
somebody
ea8df4c0d3
Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv
2023-04-14 20:38:56 -05:00
somebody
38c53191d3
possible fix for cache dl thing
2023-04-14 20:25:03 -05:00
ebolam
be0319cf03
added VERSION link in the left flyout menu footer that will show which git REPO and BRANCH the system is running on.
...
Fix for pasting text into the game area
Fix for ctrl+a then delete
Fix for changing text then back/forward a bunch loosing text
2023-04-14 13:54:11 -04:00
Henk
bde9c6980f
Transformers 4.28 support
2023-04-14 14:13:46 +02:00
henk717
03ab4b25af
Merge branch 'KoboldAI:main' into united
2023-04-14 13:59:55 +02:00
biscober
35f908e147
Update install_requirements.bat ( #7 )
...
* Update install_requirements.bat
move command to dismount temp B drive to after pip install command which requires B drive to still be mounted
* Update install_requirements.bat
cmd /k not necessary
* Update install_requirements.bat
add quotes (probably not required but w/e)
2023-04-11 04:37:48 +02:00
0cc4m
687d107d20
Update README, remove steps that are no longer required
2023-04-10 22:46:12 +02:00
0cc4m
b628aec719
Automatic installation of the quant_cuda module during install_requirements
...
Kepler (K40+) and Maxwell support
2023-04-10 22:37:16 +02:00
henk717
2385a34098
Merge pull request #325 from YellowRoseCx/patch-1
...
Add IP Whitelisting to --host
2023-04-10 14:08:09 +02:00
somebody
334c09606b
Fix for tokenizer stuff on pythia
2023-04-09 18:23:58 -05:00
somebody
3e8e3a18b0
Fix for custom gpt2
2023-04-09 18:23:52 -05:00
somebody
f73a8bb808
Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv
2023-04-09 14:38:09 -05:00
somebody
fedbffd07b
Small fixes
...
Typos galore!
2023-04-09 13:35:28 -05:00
0cc4m
7efd314428
Improve guide
2023-04-07 20:10:24 +02:00
0cc4m
ccf34a5edc
Fix merge issues with upstream, merge changes
2023-04-07 19:51:07 +02:00
0cc4m
636c4e5a52
Update gptq repo
2023-04-07 11:48:57 +02:00
YellowRoseCx
ac98cd6dd1
add IP_whitelisting to koboldai_settings.py
2023-04-05 21:27:59 -05:00
YellowRoseCx
71e5d23a5b
Add IP whitelisting to --host
2023-04-05 21:23:24 -05:00
0cc4m
40092cc9fa
Improve guide formatting
2023-04-05 21:49:13 +02:00
0cc4m
8b4375307c
Update file formatting section in guide
2023-04-05 21:10:40 +02:00
0cc4m
e4f8a9344c
Merge pull request #1 from Digitous/patch-1
...
Add install instructions
2023-04-05 21:08:14 +02:00
Henk
80e4b9e536
Merge branch 'main' into united
2023-04-05 00:22:30 +02:00
henk717
29c2d4b7a6
Removing Pygmalion from the TPU colab to get it unbanned
2023-04-04 19:51:18 +02:00
henk717
fd12214091
Clean the description of the GPU colab
2023-04-04 19:40:22 +02:00
henk717
bb51127bbf
We no longer support Pygmalion on Colab due to Google's Pygmalion ban
2023-04-04 19:37:15 +02:00
Henk
4b71da1714
Horde settings in the UI
2023-04-04 17:20:43 +02:00
0cc4m
ce6761e744
Fix issue causing expected scalar type Float but found Half RuntimeErrors
2023-04-04 07:46:53 +02:00
Henk
8bf533da9a
Pin Accelerate Version
2023-04-04 01:47:59 +02:00
somebody
8412f83ce5
Breakmodel: Fix typo
2023-04-03 18:41:18 -05:00
0cc4m
b9df9b6f59
Improve CPU offloading speed significantly when offloading less than half of the layers
2023-04-03 20:27:17 +02:00
0cc4m
5abdecad2c
Merge pull request #5 from 0cc4m/cpu-offload-1
...
CPU Offloading Support
2023-04-03 06:52:48 +02:00
0cc4m
ec4177a6d6
Remove cudatoolkit-dev and gcc/gxx 9 from conda env because they didn't resolve on Windows
2023-04-03 06:50:36 +02:00
somebody
4230fe4229
Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv
2023-04-02 16:41:21 -05:00
somebody
77f0797b1a
Model fix
2023-04-02 15:47:52 -05:00
somebody
9d70646e4d
Lazyload: Safetensors
2023-04-02 15:40:34 -05:00
0cc4m
c8d00b7a10
Add CPU offloading support for GPT-NeoX, GPT-J and OPT
2023-04-02 18:36:31 +02:00
0cc4m
e742083703
Fix multi-gpu-offloading
2023-04-02 11:17:29 +02:00
0cc4m
2729b77640
Add offload.py adapted from llama_inference_offload.py, with multi-gpu support and some improvements. Not yet functional, and still just supports Llama
2023-04-02 10:32:19 +02:00
Henk
4a8b099888
Model loading fix
2023-04-02 00:29:56 +02:00
0cc4m
110f8229c5
Add cudatoolkit-dev for compilation, compatible gcc 9 and update transformers to fix error in gptq
2023-04-01 21:33:05 +02:00