somebody
b256a8fbc7
Debug
2023-04-27 19:33:03 -05:00
onesome
467f2f25eb
More loading fixes
2023-04-26 16:58:33 -05:00
onesome
d4f7b60dc9
Fix for multiple paths
2023-04-26 16:49:12 -05:00
onesome
6776a71532
Add more info to custom model error
2023-04-26 16:36:52 -05:00
onesome
bbf4963d6e
Fix custmodpth stuff for hf loading
2023-04-26 16:18:45 -05:00
onesome
c146ae9d84
Delete legacy gpt2 custom loader
2023-04-26 16:07:18 -05:00
onesome
9579298df7
Better fallback
2023-04-25 22:28:07 -05:00
onesome
6e3aebc1ea
Zap debug
2023-04-25 21:13:17 -05:00
onesome
d496e861f4
Undo pretty code because I haven't cracked the jax enigma yet
2023-04-25 21:11:49 -05:00
onesome
1db9d9ba61
Lazyload: Whoops
2023-04-25 18:46:54 -05:00
onesome
e28e268a2d
Use safetensors only when available
2023-04-25 18:32:37 -05:00
onesome
0268305cfe
Change fallback notifications to warnings
2023-04-25 18:26:49 -05:00
onesome
b8bef641ff
Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv
2023-04-25 16:54:53 -05:00
somebody
f9fb5eba89
Remove debug
2023-04-15 18:56:49 -05:00
somebody
5dd67d027a
Workaround for socketio context errors for loading
2023-04-15 18:54:21 -05:00
somebody
08b4e317ff
Fix double slashing
2023-04-15 13:30:05 -05:00
somebody
d3a73aaeba
Fix api
2023-04-15 13:17:20 -05:00
somebody
4dcf570407
Fix legacy model loading
2023-04-15 12:57:35 -05:00
one-some
1b500c7179
Merge pull request #5 from LostRuins/concedo_api
...
Added stop sequences functionality for API calls
2023-04-15 10:51:31 -05:00
somebody
2b950f08d3
Remove legacy no accelerate fallback code
...
Was causing issues with disk cache the old code had a
`and not utils.HAS_ACCELERATE` preceding it (a variable which no longer
exists), and since disk cache is accelerate only, there was no disk
handling code in here. Anyway its bad so blast it
2023-04-15 10:47:31 -05:00
Concedo
9705b7b79c
increase API version (+1 squashed commits)
...
Squashed commits:
[c168c08
] Added stop sequences functionality for API calls
2023-04-15 18:09:53 +08:00
somebody
ea8df4c0d3
Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv
2023-04-14 20:38:56 -05:00
somebody
38c53191d3
possible fix for cache dl thing
2023-04-14 20:25:03 -05:00
somebody
334c09606b
Fix for tokenizer stuff on pythia
2023-04-09 18:23:58 -05:00
somebody
8412f83ce5
Breakmodel: Fix typo
2023-04-03 18:41:18 -05:00
somebody
77f0797b1a
Model fix
2023-04-02 15:47:52 -05:00
somebody
9d70646e4d
Lazyload: Safetensors
2023-04-02 15:40:34 -05:00
somebody
91bb433b5f
GenericTokenizer: Fall back to defined tokenizer
...
Shouldn't be relied on for model-agnostic code, but for loading
processes where you know the tokenizer class used it should be okie
dokie
2023-03-19 19:03:20 -05:00
somebody
ffe85ce8a1
Modeling: Fix logits processors (probs, biasing, lua)
2023-03-17 16:56:47 -05:00
somebody
692dbfeb37
Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv
2023-03-17 16:20:13 -05:00
somebody
8d0bc404a5
Model: More Jax import fixes and formatting
2023-03-17 15:36:44 -05:00
somebody
03af06638c
Modeling: Maybe fix samplers
2023-03-13 20:42:35 -05:00
somebody
b93c339145
Model: Lazyload backends
2023-03-13 20:29:29 -05:00
somebody
938c97b75a
RWKV: Fix yet another typo
2023-03-13 19:39:19 -05:00
somebody
14b2543c7c
RWKV: Fix typo
2023-03-13 19:36:58 -05:00
somebody
b10b201701
Model: Add basic RWKV implementation
2023-03-13 19:34:38 -05:00
somebody
bf8b60ac2d
Model: Add GenericTokenizer
...
Because Hugging Face doesnt have a consistant API across their own
libraries
2023-03-13 17:36:58 -05:00
somebody
0320678b27
Model: WIP horde and API tests
2023-03-13 14:11:06 -05:00
somebody
cd8ccf0a5e
Modeling: Add seed parameter to raw_generate
...
Yahooo, decoupling from koboldai_vars. This makes the generation test
pass in `test_generation.py`, and makes full determinism outside of
core_generate work.
2023-03-12 21:49:10 -05:00
somebody
d2696dad66
Remove ancient debug
2023-03-10 20:00:39 -06:00
somebody
8937e7f6df
Model: Add basic tests
...
We now do some basic tests for:
- hf torch loading (normal, lazy, lowmem)
- hf torch generation (shape batches, shape tokencount, faulty
determinism)
Currently full determinism is failing; yahoo, the tests work!
All of the tests initally failed (note the test environment functions
different from the aiserver environment due to aiserver doing a lot of
initalizing stuff, working on phasing that out) but now only one fails.
Very useful for finding bugs!
2023-03-10 20:00:39 -06:00
somebody
8c8bdfaf6a
Model: Fix assorted bugs
...
and ignore warnings in pytest
2023-03-10 20:00:39 -06:00
somebody
3646aa9e83
Model: Respect model lazyload over kaivars
...
kaivars dictates model config unless its from outside aiserver or
whatever.
2023-03-10 20:00:39 -06:00
somebody
52095054a3
Model: Reformat and clean up
2023-03-10 20:00:39 -06:00
somebody
fb0b2f0467
Model: Ditch awful current_model hack
...
thanks to whjms for spotting that this could be zapped
2023-03-10 20:00:38 -06:00
somebody
cb6010d666
Model: Respect sampler bounds in torch
...
A rather embarassing way to spend an hour debugging after I told myself
"I'd better remember to add this important thing to the torch side".
Samplers were being applied when in their "off values" causing
boring mathmatical operations to take place (ie anything x 0 is always
0)
2023-03-07 21:14:37 -06:00
somebody
6b45367cc7
Model: Rename "Colab" to "Basic API"
2023-03-07 17:07:01 -06:00
somebody
dc3dd38c4e
Merge branch 'united' of https://github.com/henk717/KoboldAI into model-structure-and-maybe-rwkv
2023-03-07 17:02:28 -06:00
somebody
3b05359e4b
Model: Refuse to serve certain models over the API
2023-03-07 16:54:08 -06:00
somebody
3a128e76b4
Attempts at dynamic wi fixes
2023-03-07 16:33:23 -06:00