Henk
cf4adfc7b4
No longer mess with quotes
2023-09-06 04:50:58 +02:00
Henk
ac61cd8e59
Slow tokenizer first
2023-09-06 04:35:25 +02:00
henk717
0e7f27247f
Merge pull request #453 from AlpinDale/united
...
Add new models
2023-09-06 03:50:10 +02:00
henk717
256f633f11
Merge pull request #454 from ebolam/Model_Plugins
...
Fix for options not showing up on refresh
2023-09-06 03:43:46 +02:00
ebolam
9f1e87fb49
Remove debug text
2023-09-05 19:22:25 -04:00
ebolam
fb77d8ebd5
Merge branch 'Model_Plugins' of https://github.com/ebolam/KoboldAI into Model_Plugins
2023-09-05 19:21:44 -04:00
ebolam
bcfe5c7fdd
Fix for text options not showing after multi-gen when refreshing web page.
2023-09-05 19:21:31 -04:00
ebolam
2d0040871d
Fix for text options not showing after multi-gen when refreshing web page.
2023-09-05 19:20:35 -04:00
AlpinDale
9209a9fdab
Add new models
2023-09-05 20:07:08 +00:00
henk717
4260fa60ad
Merge pull request #451 from ebolam/Model_Plugins
...
Fix for mobile touch on setting sliders
2023-09-04 22:31:37 +02:00
ebolam
0a55d9c735
Merge branch 'Model_Plugins' of https://github.com/ebolam/KoboldAI into Model_Plugins
2023-09-04 16:03:04 -04:00
ebolam
45520684df
Fix for mobile (Tested on Chrome Android Samsung phone)
2023-09-04 16:03:00 -04:00
ebolam
e2eb96f443
Merge branch 'henk717:united' into Model_Plugins
2023-09-04 15:32:12 -04:00
ebolam
5d23f10bf3
Fix for mobile touch
2023-09-04 15:31:30 -04:00
henk717
5c6ff8faa5
Merge pull request #450 from Disty0/united
...
Initial Intel ARC support with IPEX
2023-09-04 20:56:34 +02:00
Disty0
7cd23ed1a8
Fix return_xpu
2023-09-04 11:43:59 +03:00
Disty0
8151948698
Initial Intel ARC support with IPEX
2023-09-04 04:06:41 +03:00
Henk
af330f0445
Remove legnth cap
2023-09-04 01:23:31 +02:00
Henk
cc213e028d
Ignore unknown generation parameters
2023-09-04 00:41:39 +02:00
Henk
0ab11e1322
Allow unlimited stop sequences
2023-09-03 23:25:35 +02:00
henk717
ce32dddef3
Merge pull request #449 from LostRuins/concedo_united_PR
...
Updated Kobold Lite to v60
2023-09-03 16:09:06 +02:00
Concedo
966b6f8a2e
Updated Kobold Lite to v60
2023-09-03 11:40:28 +08:00
Henk
4c03402e33
API Example Polish and Fixes
2023-09-02 03:41:01 +02:00
henk717
d1430d703d
Merge pull request #448 from YellowRoseCx/patch-3
...
Create api_example.py
2023-09-02 03:39:25 +02:00
YellowRoseCx
cb577bd491
Create api_example.py
2023-09-01 20:38:58 -05:00
Henk
565ab8a38f
badwordids -> badwordsids typofix
2023-08-30 17:00:51 +02:00
henk717
1d32d16ba4
Merge branch 'KoboldAI:main' into united
2023-08-30 13:13:20 +02:00
Llama
fe53cb2f7e
Merge pull request #70 from pi6am/feat/exllama-unban-eos
...
Hook up use_default_badwordids in exllama
2023-08-29 23:17:37 -07:00
Llama
d6ed75f993
Hook up use_default_badwordids in exllama
...
Use the value of the use_default_badwordids setting to configure
bad_words_ids. Also add square brackets to bad_words_ids if the
use_default_badwordids setting is True. Fix an issue with
attempting to use the tokenizer too early, and fix an exception
populating Lua bridge data when zero tokens are generated, which
can now happen if use_default_badwordids is False and the first
token generated is EOS.
2023-08-29 23:08:51 -07:00
Llama
36f53cc915
Merge pull request #69 from pi6am/merge/united-exllama
...
Merge branch henk717/united into exllama
2023-08-29 19:54:14 -07:00
Llama
b5b0e3faea
Merge branch 'henk717/united' into merge/united-exllama
2023-08-29 19:48:44 -07:00
Henk
49fa63052f
Allow EOS unbanning
2023-08-29 20:51:09 +02:00
Henk
d77acf17eb
Transformers 4.32.1
2023-08-28 22:55:33 +02:00
Henk
ec4cecf2f8
Restore Horde UI for local users
2023-08-28 21:46:30 +02:00
Henk
c29a5019b3
Disable Horde UI due to lockups
2023-08-28 20:02:29 +02:00
Llama
6e64763b63
Merge pull request #68 from pi6am/merge/united-exllama
...
Merge henk717/united into exllama
2023-08-28 09:57:35 -07:00
Llama
2c48e05f7c
Add exllama dependency back to requirements.
2023-08-28 09:52:31 -07:00
Llama
6151cbd053
Merge branch 'united' into merge/united-exllama
2023-08-28 09:32:19 -07:00
henk717
030ca26aa3
Merge pull request #445 from db0/scribe
...
fix: Avoid conflicting args
2023-08-28 18:04:15 +02:00
db0
b20b910568
remove debug
2023-08-28 17:49:34 +02:00
db0
31161409f6
fix argsv placement
2023-08-28 17:47:52 +02:00
db0
25b78abcd2
Merge branch 'united' into scribe
2023-08-28 17:45:49 +02:00
Llama
5229987ab7
Merge pull request #66 from pi6am/feat/exllama-config
...
Modify exllama to load unrenamed gptq quantized models
2023-08-28 00:09:50 -07:00
Llama
554af7b175
Modify exllama to load unrenamed gptq quantized models
...
Read config.json and enable exllama loading if the model has a
`quantization_config` with `quant_methdod` of `gptq`. Note that this
implementation is limited and only supports model.safetensors.
That said, this supports loading popular gptq quantized models
without renaming or symlinking the model file.
2023-08-27 23:56:02 -07:00
Llama
812df5ea56
Merge pull request #65 from pi6am/feat/exllama-badwords
...
Add the eos token to exllama bad words.
2023-08-27 17:03:25 -07:00
Llama
08ff7c138c
Add the eos token to exllama bad words.
...
The bos token was already hardcoded as a bad word id.
Store badwords in a list and iterate over them during generation.
Add the Llama eos token to the list of bad words.
Also support "single line mode", which adds newline (13) to badwords.
2023-08-27 16:34:52 -07:00
Henk
3e0b8279f2
Rename GPTQ loading
2023-08-27 20:51:14 +02:00
Llama
0d150e412e
Merge pull request #64 from pi6am/fix/multinomial-workaround
...
Resample to work around a bug in torch.multinomial
2023-08-26 22:42:21 -07:00
Llama
b7e38b4757
Resample to work around a bug in torch.multinomial
...
There is a bug in PyTorch 2.0.1 that allows torch.multinomial to
sometimes choose elements that have zero probability. Since
this is uncommon we can continue to use torch.multinomial as
long as we verify that the results are valid. If they aren't,
try again until the probability of each selected token is positive.
2023-08-26 22:26:26 -07:00
Henk
290f2ce05e
CPU only warning
2023-08-26 00:03:28 +02:00