Commit Graph

966 Commits

Author SHA1 Message Date
072ca87977 Load soft prompt at the end instead of inside loadsettings() 2022-02-23 21:15:08 -05:00
8120e4dfa2 Need to set vars.allowsp to True before calling loadsettings() 2022-02-23 21:09:31 -05:00
c45ba497c9 Load settings earlier to avoid TPU badwords issues 2022-02-23 20:39:11 -05:00
ac59e55d62 Smaller optimizations 2022-02-24 01:14:26 +01:00
8e9d9faa97 Merge pull request #82 from VE-FORBRYDERNE/tpu-config
Allow TPU models to specify settings/config in config.json
2022-02-24 00:53:40 +01:00
ad10ac8871 Allow TPU models to specify settings/config in config.json 2022-02-23 18:22:18 -05:00
7de3311000 Fix sentencepiece model saving 2022-02-23 22:04:41 +01:00
6151d16df0 Merge pull request #81 from VE-FORBRYDERNE/dematerialized
Use dematerialized loading in TPU backend for lower device memory usage
2022-02-23 07:11:26 +01:00
7ec549c726 Use dematerialized loading in TPU backend for lower device memory usage 2022-02-22 19:43:13 -05:00
fd7ba9f70e Also check for Config in models/ 2022-02-22 19:22:08 +01:00
306d96a8eb Seperate Drive Disconnect 2022-02-22 18:03:06 +01:00
a0518edc36 Temporary Transformers Git for XGLM 2022-02-22 02:42:04 +01:00
74012a24c9 Expose GDrive Models 2022-02-22 02:35:27 +01:00
9aeae94d0e Cleanup leakage (Didn't appear in my commit list) 2022-02-22 02:32:02 +01:00
cb6ccacd64 Dependencies required for newer models 2022-02-21 21:17:12 +01:00
4ace11f5b8 Merge pull request #80 from VE-FORBRYDERNE/xglm-position-ids
Temporary fix for XGLM positional embedding issues
2022-02-21 00:47:20 +01:00
300db651de Open models folder by default 2022-02-21 00:46:18 +01:00
da10e2dc1d Don't crash if XGLMSinusoidalPositionalEmbedding doesn't exist 2022-02-20 17:41:00 -05:00
5dc4969173 Temporary fix for XGLM positional embedding issues 2022-02-20 14:17:24 -05:00
7c678820cd Exclude Models from our Git 2022-02-20 19:36:14 +01:00
27cf59bb94 Merge pull request #79 from VE-FORBRYDERNE/xglm-eos
Prevent transformers XGLM from stopping generation on `</s>` token
2022-02-20 19:03:51 +01:00
a63fa3b067 Prevent transformers XGLM from stopping generation on </s> token 2022-02-19 23:15:16 -05:00
70e0295600 Merge branch 'KoboldAI:main' into united 2022-02-19 23:34:46 +01:00
acc5804820 Merge pull request #97 from mrseeker/patch-2
Add description of Janeway
2022-02-19 23:27:13 +01:00
ba7f0de0d5 Merge pull request #98 from AngryBeeSec/main
Update play-cuda.sh
2022-02-19 23:22:04 +01:00
b6d885cf0a Update play-cuda.sh
Allows the use of newer models
2022-02-19 16:26:20 -05:00
a47e93cee7 Seperate Low Memory Mode
In 1.16 we had significantly faster loading speeds because we did not do as much memory conservation, its time to give users the choice. If you want the original faster behavior and have the memory run KoboldAI as usual. Otherwise run play-lowmem.bat or aiserver.py with --lowmem. For colab this is still the default behavior to avoid breaking models that would otherwise load fine.
2022-02-18 16:21:28 +01:00
4c84d731db Merge branch 'KoboldAI:main' into united 2022-02-18 15:02:24 +01:00
2b133548be Add description of Janeway 2022-02-18 14:37:28 +01:00
90be138ac5 Add Janeway to the GPU Colab 2022-02-18 14:26:29 +01:00
8e03f1c612 Merge branch 'KoboldAI:main' into united 2022-02-18 14:21:34 +01:00
f06acb59be Add the Janeway model
New model released by Mr.Seeker
2022-02-18 14:18:41 +01:00
cba93e29d2 Update aiserver.py 2022-02-18 02:11:08 +01:00
76a6c124dd Quiet on Colab
Makes the Colab mode also automatically activate the Quiet mode to improve privacy. We should no longer need this in the colab console thanks to the redo feature. Need something different for testing? Use --remote instead.
2022-02-18 02:07:40 +01:00
02246dfc4d Remote play improvements
Change the proposed --share to --unblock to make it more apparent what this feature does. The feature unblocks the port from external access, but does not add remote play support. For remote play support without a proxy service I have added --host .
2022-02-18 01:08:12 +01:00
9b72583110 Merge branch 'KoboldAI:main' into united 2022-02-18 00:37:34 +01:00
e571a17f84 Update readme.md 2022-02-15 20:18:21 +01:00
a05aef552c Merge branch 'KoboldAI:main' into united 2022-02-14 18:10:56 +01:00
ca5b9f968f Merge pull request #76 from VE-FORBRYDERNE/newline
Fix fairseq newline handling issues
2022-02-14 18:10:25 +01:00
50a96485a9 Fix dm-haiku
They did a change that breaks compatibility with our other dependencies, forcing version 0.0.5 to fix this.
2022-02-14 18:05:50 +01:00
ec54bc9d9b Fix typo in send_debug() 2022-02-12 20:11:35 -05:00
f682c1229a Fix fairseq newline handling issues 2022-02-12 13:23:59 -05:00
c1af8f72c3 Merge pull request #75 from ebolam/united
Fixed retry bug due to redo/pin code
2022-02-11 03:27:51 +01:00
633152ee84 Fixed Retry bug due to redo/pin code 2022-02-10 10:01:07 -05:00
cd00373cfb Deleted unused svg 2022-02-10 09:21:07 -05:00
e1ef4e4fa8 Merge pull request #74 from ebolam/united
Redo, Pinning, and docker enhancements
2022-02-07 01:06:36 +01:00
c0bbe9f810 Reverted docker-cuda to mainline version. 2022-02-06 19:04:13 -05:00
586b989582 Redo bug fix 2022-02-06 18:53:24 -05:00
98609a8abc Merge branch 'united' of https://github.com/ebolam/KoboldAI into united 2022-02-06 13:48:34 -05:00
80ae054cb5 Merge branch 'henk717:united' into united 2022-02-06 13:42:59 -05:00