somebody
aca2b532d7
Remove debug
2023-06-21 14:15:38 -05:00
somebody
5f224e1366
Restore choice of lazyload or not
2023-06-21 14:13:14 -05:00
somebody
0052ad401a
Basic breakmodel ui support
...
Seems to work
2023-06-21 13:57:32 -05:00
somebody
f326fc07e8
Seems to work
2023-05-31 14:42:05 -05:00
somebody
24b0b32829
Maybe works now...?
2023-05-31 14:31:08 -05:00
somebody
ac4384ef75
Auto _no_split_modules
2023-05-31 10:55:46 -05:00
somebody
58ffad237b
OPT hack
2023-05-29 13:34:11 -05:00
somebody
ceaefa9f5e
Not quite
2023-05-28 14:57:45 -05:00
somebody
1546b9efaa
Hello its breaking breakmodel time
2023-05-27 16:31:53 -05:00
ebolam
5561cc1f22
Fix for GPU generation
2023-05-23 08:33:19 -04:00
ebolam
4c25d6fbbb
Fix for loading model multiple times loosing the gpu/cpu splits
2023-05-22 20:34:01 -04:00
ebolam
9e53bcf676
Fix for breakmodel loading to CPU when set to GPU
2023-05-22 20:24:57 -04:00
ebolam
f1a16f260f
Potential breakmodel fix
2023-05-22 16:10:41 -04:00
ebolam
ca770844b0
Fix for breakmodel
2023-05-22 15:07:59 -04:00
ebolam
3db231562f
Merge pull request #382 from henk717/united
...
Update to united
2023-05-19 06:05:25 -04:00
ebolam
56d2705f4b
removed breakmodel command line arguments (except nobreakmodel)
2023-05-18 20:19:33 -04:00
ebolam
06f59a7b7b
Moved model backends to separate folders
...
added some model backend settings save/load
2023-05-18 20:14:33 -04:00
Henk
205c64f1ea
More universal pytorch folder detection
2023-05-13 20:26:55 +02:00
somebody
3065c1b40e
Ignore missing keys in get_original_key
2023-05-11 17:10:43 -05:00
somebody
c16336f646
Add traceback to debug log on fallback
2023-05-11 17:10:19 -05:00
ebolam
71aee4dbd8
First concept of model plugins with a conceptual UI.
...
Completely breaks UI2 model loading.
2023-05-10 16:30:46 -04:00
Henk
bb206f598e
Don't load peft when unused
2023-05-06 18:55:26 +02:00
somebody
b7db709c47
PEFT: Change directory structure to be inside model
2023-05-06 11:16:09 -05:00
somebody
f02ddab7c7
Merge branch 'united' of https://github.com/henk717/KoboldAI into peft
2023-05-06 10:47:14 -05:00
Henk
33969b5845
Basic HF code execution support
2023-05-05 17:23:01 +02:00
somebody
35b56117e6
Basic PEFT support
2023-05-03 18:51:01 -05:00
somebody
efe268df60
Move overrides to better places
2023-05-02 20:18:33 -05:00
somebody
f6b5548131
Support safetensors in get_sharded_checkpoint_num_tensors
2023-05-01 19:15:27 -05:00
somebody
97e84928ba
Download all shards correctly on aria2 and raise on bad load key
2023-05-01 18:53:36 -05:00
somebody
933dbd634a
HFInferenceModel: Make badwordsids not unique to torch
2023-05-01 17:13:33 -05:00
onesome
467f2f25eb
More loading fixes
2023-04-26 16:58:33 -05:00
onesome
9579298df7
Better fallback
2023-04-25 22:28:07 -05:00
onesome
b8bef641ff
Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv
2023-04-25 16:54:53 -05:00
somebody
f9fb5eba89
Remove debug
2023-04-15 18:56:49 -05:00
somebody
5dd67d027a
Workaround for socketio context errors for loading
2023-04-15 18:54:21 -05:00
one-some
1b500c7179
Merge pull request #5 from LostRuins/concedo_api
...
Added stop sequences functionality for API calls
2023-04-15 10:51:31 -05:00
somebody
2b950f08d3
Remove legacy no accelerate fallback code
...
Was causing issues with disk cache the old code had a
`and not utils.HAS_ACCELERATE` preceding it (a variable which no longer
exists), and since disk cache is accelerate only, there was no disk
handling code in here. Anyway its bad so blast it
2023-04-15 10:47:31 -05:00
Concedo
9705b7b79c
increase API version (+1 squashed commits)
...
Squashed commits:
[c168c08
] Added stop sequences functionality for API calls
2023-04-15 18:09:53 +08:00
somebody
ea8df4c0d3
Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv
2023-04-14 20:38:56 -05:00
somebody
77f0797b1a
Model fix
2023-04-02 15:47:52 -05:00
somebody
9d70646e4d
Lazyload: Safetensors
2023-04-02 15:40:34 -05:00
somebody
ffe85ce8a1
Modeling: Fix logits processors (probs, biasing, lua)
2023-03-17 16:56:47 -05:00
somebody
692dbfeb37
Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv
2023-03-17 16:20:13 -05:00
somebody
0320678b27
Model: WIP horde and API tests
2023-03-13 14:11:06 -05:00
somebody
cd8ccf0a5e
Modeling: Add seed parameter to raw_generate
...
Yahooo, decoupling from koboldai_vars. This makes the generation test
pass in `test_generation.py`, and makes full determinism outside of
core_generate work.
2023-03-12 21:49:10 -05:00
somebody
8c8bdfaf6a
Model: Fix assorted bugs
...
and ignore warnings in pytest
2023-03-10 20:00:39 -06:00
somebody
3646aa9e83
Model: Respect model lazyload over kaivars
...
kaivars dictates model config unless its from outside aiserver or
whatever.
2023-03-10 20:00:39 -06:00
somebody
52095054a3
Model: Reformat and clean up
2023-03-10 20:00:39 -06:00
somebody
fb0b2f0467
Model: Ditch awful current_model hack
...
thanks to whjms for spotting that this could be zapped
2023-03-10 20:00:38 -06:00
somebody
cb6010d666
Model: Respect sampler bounds in torch
...
A rather embarassing way to spend an hour debugging after I told myself
"I'd better remember to add this important thing to the torch side".
Samplers were being applied when in their "off values" causing
boring mathmatical operations to take place (ie anything x 0 is always
0)
2023-03-07 21:14:37 -06:00