Commit Graph

1618 Commits

Author SHA1 Message Date
d94f29a68a Merge pull request #127 from VE-FORBRYDERNE/tracer
Fix JAX UnexpectedTracerError
2022-06-23 19:29:51 +02:00
8098f4ec8f Merge branch 'KoboldAI:main' into united 2022-06-23 17:20:48 +02:00
1d41966d88 Merge pull request #129 from VE-FORBRYDERNE/budget
Account for lnheader in budget calculation
2022-06-23 12:39:20 +02:00
0eb9f8a879 Account for lnheader in budget calculation 2022-06-22 19:16:24 -04:00
3de22f2b27 Merge pull request #160 from VE-FORBRYDERNE/gc
Delete all torch tensors before loading model
2022-06-22 18:38:21 +02:00
53034ee533 Delete all torch tensors before loading model 2022-06-22 12:07:36 -04:00
f127918114 Merge pull request #159 from VE-FORBRYDERNE/fairseq
Don't blacklist </s> token in "s" newline mode
2022-06-22 17:41:20 +02:00
922394c68f Don't blacklist </s> token in "s" newline mode 2022-06-22 11:23:03 -04:00
d4e18360f0 HF NeoX Support 2022-06-22 01:46:40 +02:00
5f9a116052 Merge pull request #128 from VE-FORBRYDERNE/neox
TPU support for HF GPT-NeoX model
2022-06-22 01:39:53 +02:00
8c594c6869 Correct the padding token for GPT-NeoX 2022-06-21 19:37:43 -04:00
a7f667c34c Use NeoX badwords when loading from HF GPT-NeoX model 2022-06-21 19:33:25 -04:00
5e3c7c07ae Merge branch 'main' into neox 2022-06-21 19:30:51 -04:00
f1d0a327f8 Merge branch 'KoboldAI:main' into united 2022-06-21 23:34:32 +02:00
75bc472a9f Transformers bump to 4.20.1
Transformers issued an important change for the OPT models breaking their compatibility with all older versions. In order for people to be able to use all models on the menu they need 4.20.1 so this is now forced in the dependencies making the update easier.
2022-06-21 23:33:38 +02:00
b5b8e5a30b Merge branch 'KoboldAI:main' into united 2022-06-21 23:19:57 +02:00
2be1f5088f Merge pull request #126 from VE-FORBRYDERNE/opt
Update OPT models and fix 20B model on TPU
2022-06-21 23:19:03 +02:00
33a2a318db Fix 20B TPU model 2022-06-21 17:16:01 -04:00
a7e3ef71aa Add final layer norm to OPT 2022-06-21 16:36:26 -04:00
37eb47d0d3 Merge pull request #157 from VE-FORBRYDERNE/sp-fix
Bug fixes and new soft prompt implementation
2022-06-21 22:20:36 +02:00
8593bf339b Another typo fix 2022-06-21 15:36:25 -04:00
7e0ded6b47 Typo fix 2022-06-21 15:12:55 -04:00
91643be10a Change soft prompt implementation to a more universal one 2022-06-21 15:03:43 -04:00
0ea4fa9c87 Automatically calculate badwords and pad_token_id 2022-06-21 14:35:52 -04:00
ea7d278ff4 Fix 20B TPU model 2022-06-21 13:16:45 -04:00
6b172306f6 move_model_to_devices no longer crashes if you don't have accelerate 2022-06-21 13:15:46 -04:00
f2c5bb5cb7 Merge pull request #156 from VE-FORBRYDERNE/accelerate
Accelerate disk cache support
2022-06-21 00:31:50 +02:00
ff69e9fbfe Put layers_module_names, module_names and named_buffers in utils.py 2022-06-20 17:17:42 -04:00
1620ac4148 Lazy loader needs to cache named buffers of layers in the disk cache 2022-06-20 17:08:52 -04:00
ab5ab79003 Set primary device to CPU if in CPU-only mode 2022-06-20 16:25:01 -04:00
bd7d7b41a1 Don't enable accelerate if no layers are in disk cache or GPUs 2022-06-20 16:21:44 -04:00
90fd8b1845 Disk cache support in CPU-only mode 2022-06-20 16:06:09 -04:00
af07d7a15f Disk cache support for computers with at least one GPU 2022-06-20 14:49:54 -04:00
47a58a36b8 Add disk cache slider 2022-06-19 22:53:30 -04:00
efed44ac8d Merge pull request #155 from VE-FORBRYDERNE/accelerate
Initial support for Accelerate
2022-06-20 01:08:54 +02:00
4dd59e0a9d Correct the type hint for lazy_load_callback 2022-06-19 17:17:41 -04:00
21de36c4b0 Lazy loader now moves all non-layer weights to primary device 2022-06-19 16:44:23 -04:00
26c319519e Lazy loader now attempts to pin layers if accelerate is enabled 2022-06-19 16:35:23 -04:00
042cf3e560 Automatically support soft prompts for all transformers models 2022-06-19 13:11:58 -04:00
cc56718a7e Fix lazy loader putting too many layers on CPU 2022-06-19 00:29:35 -04:00
1380eb0bb0 Disable lazy loader when using GPT-2 2022-06-18 23:54:11 -04:00
f9732eb143 Always enable breakmodel if accelerate is available 2022-06-18 23:46:09 -04:00
8b4efc5d0a Use accelerate.dispatch_model() instead of breakmodel if possible 2022-06-18 23:41:36 -04:00
f7ffdd7b6b Add more model querying utilities 2022-06-18 18:16:56 -04:00
e143963161 Merge branch 'united' into accelerate 2022-06-18 13:47:38 -04:00
b209cf9868 NS mode as default
Experimental change that makes NS the default, more and more models seem to be requiring this as megatron based models are getting traction, neither does this seem to break the original models (with the exception of a user not being able to use </s> in generated outputs, the extremely rare case someone would be effected by this they can manually switch the mode by editing their settings file).

If this breaks nothing ns will remain the default, however the n mode should remain a choice for those who need it. In case it does get reversed I have also added the bloom model type to the ns list since its models require this.
2022-06-18 19:46:16 +02:00
23aae24f8e Merge pull request #154 from VE-FORBRYDERNE/united-merge
Merge main into united
2022-06-18 19:42:26 +02:00
0eedc541c8 Merge branch 'main' into united-merge 2022-06-18 13:39:23 -04:00
a10446f258 Merge pull request #123 from VE-FORBRYDERNE/tokenizer
Fix OPT tokenization problems
2022-06-18 11:38:14 +02:00
5e71f7fe97 Use slow tokenizer if fast tokenizer is not available 2022-06-17 21:08:37 -04:00