somebody
|
ad4528b5a6
|
critical change
|
2023-07-24 17:17:57 -05:00 |
|
somebody
|
a73420c49c
|
really really really sketchy breakmodel implementation
im gonna go lie down for an extended period of time
|
2023-07-24 17:15:59 -05:00 |
|
somebody
|
929917efe9
|
Remove shrieking
|
2023-07-24 13:09:43 -05:00 |
|
somebody
|
4a6cccb002
|
Import fix
|
2023-07-24 13:09:15 -05:00 |
|
somebody
|
a6aafb2525
|
GPTQ: Patch QuantLinear to not use CPU RAM
|
2023-07-24 13:07:30 -05:00 |
|
somebody
|
1df03d9a27
|
Basic
|
2023-07-23 20:54:04 -05:00 |
|
0cc4m
|
973aea12ea
|
Only import big python modules for GPTQ once they get used
|
2023-07-23 22:07:34 +02:00 |
|
0cc4m
|
09bb1021dd
|
Fallback to transformers if hf_bleeding_edge not available
|
2023-07-23 07:16:52 +02:00 |
|
0cc4m
|
748e5ef318
|
Add sliders for exllama context size and related methods
|
2023-07-23 07:11:28 +02:00 |
|
0cc4m
|
9aa6c5fbbf
|
Merge upstream changes, fix conflict, adapt backends to changes
|
2023-07-19 06:56:09 +02:00 |
|
0cc4m
|
0001ae00ab
|
Add v2 with bias support (e.g. for Tulu-30b)
|
2023-06-12 07:18:22 +02:00 |
|
0cc4m
|
12df8220fb
|
Add gpt_bigcode support, fix 8-bit GPTQ incoherence
|
2023-06-12 07:14:36 +02:00 |
|
0cc4m
|
c82625490a
|
Rename gptq backend folder
|
2023-06-04 12:31:24 +02:00 |
|