Commit Graph

983 Commits

Author SHA1 Message Date
Gnome Ann c338b52d68 (torch_lazy_loader.py) Handle checkpoints with merged storage blocks 2022-03-02 01:02:35 -05:00
Gnome Ann 4fa4dbac50 Clean up when error is thrown in `use_lazy_torch_load` 2022-03-01 19:30:22 -05:00
Gnome Ann a0344b429c Upload torch_lazy_loader.py 2022-03-01 15:40:44 -05:00
ebolam 3f73f84b69 bug fix 2022-02-28 19:04:12 -05:00
Gnome Ann d8e99b12f1 Re-enable the editor mutation observer 2022-02-28 19:00:26 -05:00
henk717 50ad6864c9
Merge pull request #87 from ebolam/united
Debug and load story fix for actions_metadata variable
2022-02-28 16:58:49 +01:00
ebolam 6003b2369b Debug and load story fix for actions_metadata variable 2022-02-28 10:39:36 -05:00
henk717 261981da45
Merge pull request #86 from ebolam/united
Fixed error in redo action
2022-02-28 14:43:53 +01:00
ebolam 47d102635e
Merge branch 'united' into united 2022-02-28 08:37:45 -05:00
ebolam 7803fbb137 Fixed error in redo action when editing previous entries and/or editing right after a redo 2022-02-28 08:31:26 -05:00
henk717 13fe472264 Menu Polish 2022-02-28 02:47:15 +01:00
henk717 f628929401
Merge pull request #85 from VE-FORBRYDERNE/sp
Fix a bug with soft prompts when using transformers XGLM
2022-02-28 02:33:18 +01:00
henk717 4849a30d88
Merge pull request #84 from mrseeker/patch-3
Added KoboldAI/fairseq-dense-2.7B-Janeway
2022-02-28 02:33:07 +01:00
henk717 a466e13c00 Model List Support 2022-02-26 12:34:07 +01:00
Gnome Ann a22d59e191 Fix a bug with soft prompts when using transformers XGLM 2022-02-25 12:35:23 -05:00
Julius ter Pelkwijk 0a7376a711
Added KoboldAI/fairseq-dense-2.7B-Janeway
With pleasure I am introducing KoboldAI/fairseq-dense-2.7B-Janeway.
2022-02-24 09:00:56 +01:00
henk717 1fc173890e
Merge pull request #83 from VE-FORBRYDERNE/loadsettings
Load settings earlier to avoid TPU badwords issues
2022-02-24 04:24:28 +01:00
Gnome Ann 072ca87977 Load soft prompt at the end instead of inside `loadsettings()` 2022-02-23 21:15:08 -05:00
Gnome Ann 8120e4dfa2 Need to set `vars.allowsp` to True before calling `loadsettings()` 2022-02-23 21:09:31 -05:00
Gnome Ann c45ba497c9 Load settings earlier to avoid TPU badwords issues 2022-02-23 20:39:11 -05:00
henk717 ac59e55d62 Smaller optimizations 2022-02-24 01:14:26 +01:00
henk717 8e9d9faa97
Merge pull request #82 from VE-FORBRYDERNE/tpu-config
Allow TPU models to specify settings/config in config.json
2022-02-24 00:53:40 +01:00
Gnome Ann ad10ac8871 Allow TPU models to specify settings/config in config.json 2022-02-23 18:22:18 -05:00
henk717 7de3311000 Fix sentencepiece model saving 2022-02-23 22:04:41 +01:00
henk717 6151d16df0
Merge pull request #81 from VE-FORBRYDERNE/dematerialized
Use dematerialized loading in TPU backend for lower device memory usage
2022-02-23 07:11:26 +01:00
Gnome Ann 7ec549c726 Use dematerialized loading in TPU backend for lower device memory usage 2022-02-22 19:43:13 -05:00
henk717 fd7ba9f70e Also check for Config in models/ 2022-02-22 19:22:08 +01:00
henk717 306d96a8eb Seperate Drive Disconnect 2022-02-22 18:03:06 +01:00
henk717 a0518edc36 Temporary Transformers Git for XGLM 2022-02-22 02:42:04 +01:00
henk717 74012a24c9 Expose GDrive Models 2022-02-22 02:35:27 +01:00
henk717 9aeae94d0e Cleanup leakage (Didn't appear in my commit list) 2022-02-22 02:32:02 +01:00
henk717 cb6ccacd64 Dependencies required for newer models 2022-02-21 21:17:12 +01:00
henk717 4ace11f5b8
Merge pull request #80 from VE-FORBRYDERNE/xglm-position-ids
Temporary fix for XGLM positional embedding issues
2022-02-21 00:47:20 +01:00
henk717 300db651de Open models folder by default 2022-02-21 00:46:18 +01:00
Gnome Ann da10e2dc1d Don't crash if `XGLMSinusoidalPositionalEmbedding` doesn't exist 2022-02-20 17:41:00 -05:00
Gnome Ann 5dc4969173 Temporary fix for XGLM positional embedding issues 2022-02-20 14:17:24 -05:00
henk717 7c678820cd Exclude Models from our Git 2022-02-20 19:36:14 +01:00
henk717 27cf59bb94
Merge pull request #79 from VE-FORBRYDERNE/xglm-eos
Prevent transformers XGLM from stopping generation on `</s>` token
2022-02-20 19:03:51 +01:00
Gnome Ann a63fa3b067 Prevent transformers XGLM from stopping generation on `</s>` token 2022-02-19 23:15:16 -05:00
henk717 70e0295600
Merge branch 'KoboldAI:main' into united 2022-02-19 23:34:46 +01:00
henk717 acc5804820
Merge pull request #97 from mrseeker/patch-2
Add description of Janeway
2022-02-19 23:27:13 +01:00
henk717 ba7f0de0d5
Merge pull request #98 from AngryBeeSec/main
Update play-cuda.sh
2022-02-19 23:22:04 +01:00
AngryBeeSec b6d885cf0a
Update play-cuda.sh
Allows the use of newer models
2022-02-19 16:26:20 -05:00
henk717 a47e93cee7 Seperate Low Memory Mode
In 1.16 we had significantly faster loading speeds because we did not do as much memory conservation, its time to give users the choice. If you want the original faster behavior and have the memory run KoboldAI as usual. Otherwise run play-lowmem.bat or aiserver.py with --lowmem. For colab this is still the default behavior to avoid breaking models that would otherwise load fine.
2022-02-18 16:21:28 +01:00
henk717 4c84d731db
Merge branch 'KoboldAI:main' into united 2022-02-18 15:02:24 +01:00
Julius ter Pelkwijk 2b133548be
Add description of Janeway 2022-02-18 14:37:28 +01:00
henk717 90be138ac5 Add Janeway to the GPU Colab 2022-02-18 14:26:29 +01:00
henk717 8e03f1c612
Merge branch 'KoboldAI:main' into united 2022-02-18 14:21:34 +01:00
henk717 f06acb59be
Add the Janeway model
New model released by Mr.Seeker
2022-02-18 14:18:41 +01:00
henk717 cba93e29d2 Update aiserver.py 2022-02-18 02:11:08 +01:00