0cc4m
4d34f9b7de
Move 4-bit loading code to separate inference_model file
2023-04-16 14:20:13 +02:00
0cc4m
05b1d36803
Merge latestgptq branch onto one-some's model-structure-and-maybe-rwkv branch
2023-04-16 09:06:12 +02:00
somebody
f9fb5eba89
Remove debug
2023-04-15 18:56:49 -05:00
somebody
5dd67d027a
Workaround for socketio context errors for loading
2023-04-15 18:54:21 -05:00
somebody
08b4e317ff
Fix double slashing
2023-04-15 13:30:05 -05:00
somebody
d3a73aaeba
Fix api
2023-04-15 13:17:20 -05:00
somebody
4dcf570407
Fix legacy model loading
2023-04-15 12:57:35 -05:00
somebody
2a977feb3e
Merge branch 'model-structure-and-maybe-rwkv' of https://github.com/one-some/KoboldAI into model-structure-and-maybe-rwkv
2023-04-15 11:51:39 -05:00
somebody
a2ae87d1b7
Utils: Support safetensors aria2 download
2023-04-15 11:51:16 -05:00
0cc4m
fff2385173
Merge upstream changes
2023-04-15 18:35:54 +02:00
Henk
b68860b3de
Workaround to make --host work again
2023-04-15 18:31:39 +02:00
one-some
1b500c7179
Merge pull request #5 from LostRuins/concedo_api
...
Added stop sequences functionality for API calls
2023-04-15 10:51:31 -05:00
somebody
2b950f08d3
Remove legacy no accelerate fallback code
...
Was causing issues with disk cache the old code had a
`and not utils.HAS_ACCELERATE` preceding it (a variable which no longer
exists), and since disk cache is accelerate only, there was no disk
handling code in here. Anyway its bad so blast it
2023-04-15 10:47:31 -05:00
Henk
67334bd698
Pin accelerate version
2023-04-15 17:45:00 +02:00
somebody
b2e6fcfe3a
Remove line that sets disk_layers to None always
...
whoops
2023-04-15 10:41:10 -05:00
Henk
3eda7269f7
Fix incorrect host merge
2023-04-15 14:58:24 +02:00
Concedo
dd01cf1a93
Merge branch 'concedo_api' of https://github.com/LostRuins/KoboldAI into concedo_api
2023-04-15 18:10:28 +08:00
Concedo
9705b7b79c
increase API version (+1 squashed commits)
...
Squashed commits:
[c168c08
] Added stop sequences functionality for API calls
2023-04-15 18:09:53 +08:00
Concedo
d22423e4be
increase API version
2023-04-15 18:09:29 +08:00
Concedo
c168c08245
Added stop sequences functionality for API calls
2023-04-15 18:00:11 +08:00
somebody
ea8df4c0d3
Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv
2023-04-14 20:38:56 -05:00
somebody
38c53191d3
possible fix for cache dl thing
2023-04-14 20:25:03 -05:00
Henk
bde9c6980f
Transformers 4.28 support
2023-04-14 14:13:46 +02:00
henk717
03ab4b25af
Merge branch 'KoboldAI:main' into united
2023-04-14 13:59:55 +02:00
biscober
35f908e147
Update install_requirements.bat ( #7 )
...
* Update install_requirements.bat
move command to dismount temp B drive to after pip install command which requires B drive to still be mounted
* Update install_requirements.bat
cmd /k not necessary
* Update install_requirements.bat
add quotes (probably not required but w/e)
2023-04-11 04:37:48 +02:00
0cc4m
687d107d20
Update README, remove steps that are no longer required
2023-04-10 22:46:12 +02:00
0cc4m
b628aec719
Automatic installation of the quant_cuda module during install_requirements
...
Kepler (K40+) and Maxwell support
2023-04-10 22:37:16 +02:00
henk717
2385a34098
Merge pull request #325 from YellowRoseCx/patch-1
...
Add IP Whitelisting to --host
2023-04-10 14:08:09 +02:00
somebody
334c09606b
Fix for tokenizer stuff on pythia
2023-04-09 18:23:58 -05:00
somebody
3e8e3a18b0
Fix for custom gpt2
2023-04-09 18:23:52 -05:00
somebody
f73a8bb808
Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv
2023-04-09 14:38:09 -05:00
somebody
fedbffd07b
Small fixes
...
Typos galore!
2023-04-09 13:35:28 -05:00
0cc4m
7efd314428
Improve guide
2023-04-07 20:10:24 +02:00
0cc4m
ccf34a5edc
Fix merge issues with upstream, merge changes
2023-04-07 19:51:07 +02:00
0cc4m
636c4e5a52
Update gptq repo
2023-04-07 11:48:57 +02:00
YellowRoseCx
ac98cd6dd1
add IP_whitelisting to koboldai_settings.py
2023-04-05 21:27:59 -05:00
YellowRoseCx
71e5d23a5b
Add IP whitelisting to --host
2023-04-05 21:23:24 -05:00
0cc4m
40092cc9fa
Improve guide formatting
2023-04-05 21:49:13 +02:00
0cc4m
8b4375307c
Update file formatting section in guide
2023-04-05 21:10:40 +02:00
0cc4m
e4f8a9344c
Merge pull request #1 from Digitous/patch-1
...
Add install instructions
2023-04-05 21:08:14 +02:00
Henk
80e4b9e536
Merge branch 'main' into united
2023-04-05 00:22:30 +02:00
henk717
29c2d4b7a6
Removing Pygmalion from the TPU colab to get it unbanned
2023-04-04 19:51:18 +02:00
henk717
fd12214091
Clean the description of the GPU colab
2023-04-04 19:40:22 +02:00
henk717
bb51127bbf
We no longer support Pygmalion on Colab due to Google's Pygmalion ban
2023-04-04 19:37:15 +02:00
Henk
4b71da1714
Horde settings in the UI
2023-04-04 17:20:43 +02:00
0cc4m
ce6761e744
Fix issue causing expected scalar type Float but found Half RuntimeErrors
2023-04-04 07:46:53 +02:00
Henk
8bf533da9a
Pin Accelerate Version
2023-04-04 01:47:59 +02:00
somebody
8412f83ce5
Breakmodel: Fix typo
2023-04-03 18:41:18 -05:00
0cc4m
b9df9b6f59
Improve CPU offloading speed significantly when offloading less than half of the layers
2023-04-03 20:27:17 +02:00
0cc4m
5abdecad2c
Merge pull request #5 from 0cc4m/cpu-offload-1
...
CPU Offloading Support
2023-04-03 06:52:48 +02:00