Revert bad AutoGPTQ but keep fixes

This commit is contained in:
Henk
2023-11-08 19:18:34 +01:00
parent 9e275de5d9
commit 71cea76ae0
4 changed files with 29 additions and 24 deletions

View File

@@ -35,11 +35,11 @@ dependencies:
- flask-cors - flask-cors
- Werkzeug==2.3.7 - Werkzeug==2.3.7
- lupa==1.10 - lupa==1.10
- transformers[sentencepiece]==4.35.0 - transformers[sentencepiece]==4.34.0
- huggingface_hub==0.16.4 - huggingface_hub==0.16.4
- optimum[onnxruntime]==1.14.0 - optimum[onnxruntime]==1.13.2
- safetensors==0.4.0 - safetensors==0.3.3
- accelerate==0.24.1 - accelerate==0.21.0
- git+https://github.com/VE-FORBRYDERNE/mkultra - git+https://github.com/VE-FORBRYDERNE/mkultra
- flask-session - flask-session
- ansi2html - ansi2html
@@ -53,9 +53,10 @@ dependencies:
- git+https://github.com/0cc4m/hf_bleeding_edge/ - git+https://github.com/0cc4m/hf_bleeding_edge/
- https://github.com/0cc4m/GPTQ-for-LLaMa/releases/download/0.0.6/gptq_koboldai-0.0.6-cp38-cp38-linux_x86_64.whl; sys_platform == 'linux' - https://github.com/0cc4m/GPTQ-for-LLaMa/releases/download/0.0.6/gptq_koboldai-0.0.6-cp38-cp38-linux_x86_64.whl; sys_platform == 'linux'
- https://github.com/0cc4m/GPTQ-for-LLaMa/releases/download/0.0.6/gptq_koboldai-0.0.6-cp38-cp38-win_amd64.whl; sys_platform == 'win32' - https://github.com/0cc4m/GPTQ-for-LLaMa/releases/download/0.0.6/gptq_koboldai-0.0.6-cp38-cp38-win_amd64.whl; sys_platform == 'win32'
- auto_gptq==0.5.0 - https://github.com/PanQiWei/AutoGPTQ/releases/download/v0.4.2/auto_gptq-0.4.2+cu118-cp38-cp38-linux_x86_64.whl; sys_platform == 'linux'
- https://github.com/PanQiWei/AutoGPTQ/releases/download/v0.4.2/auto_gptq-0.4.2+cu118-cp38-cp38-win_amd64.whl; sys_platform == 'win32'
- einops - einops
- peft==0.6.0 - peft==0.3.0
- scipy - scipy
- https://github.com/0cc4m/exllama/releases/download/0.0.7/exllama-0.0.7-cp38-cp38-linux_x86_64.whl; sys_platform == 'linux' - https://github.com/0cc4m/exllama/releases/download/0.0.7/exllama-0.0.7-cp38-cp38-linux_x86_64.whl; sys_platform == 'linux'
- https://github.com/0cc4m/exllama/releases/download/0.0.7/exllama-0.0.7-cp38-cp38-win_amd64.whl; sys_platform == 'win32' - https://github.com/0cc4m/exllama/releases/download/0.0.7/exllama-0.0.7-cp38-cp38-win_amd64.whl; sys_platform == 'win32'

View File

@@ -35,11 +35,11 @@ dependencies:
- flask-cors - flask-cors
- Werkzeug==2.3.7 - Werkzeug==2.3.7
- lupa==1.10 - lupa==1.10
- transformers[sentencepiece]==4.35.0 - transformers[sentencepiece]==4.34.0
- huggingface_hub==0.16.4 - huggingface_hub==0.16.4
- optimum[onnxruntime,openvino,nncf,neural-compressor]==1.14.0 - optimum[onnxruntime,openvino,nncf,neural-compressor]==1.13.2
- safetensors==0.4.0 - safetensors==0.3.3
- accelerate==0.24.1 - accelerate==0.21.0
- git+https://github.com/VE-FORBRYDERNE/mkultra - git+https://github.com/VE-FORBRYDERNE/mkultra
- flask-session - flask-session
- ansi2html - ansi2html
@@ -51,9 +51,10 @@ dependencies:
- git+https://github.com/0cc4m/hf_bleeding_edge/ - git+https://github.com/0cc4m/hf_bleeding_edge/
- https://github.com/0cc4m/GPTQ-for-LLaMa/releases/download/0.0.6/gptq_koboldai-0.0.6-cp38-cp38-linux_x86_64.whl; sys_platform == 'linux' - https://github.com/0cc4m/GPTQ-for-LLaMa/releases/download/0.0.6/gptq_koboldai-0.0.6-cp38-cp38-linux_x86_64.whl; sys_platform == 'linux'
- https://github.com/0cc4m/GPTQ-for-LLaMa/releases/download/0.0.6/gptq_koboldai-0.0.6-cp38-cp38-win_amd64.whl; sys_platform == 'win32' - https://github.com/0cc4m/GPTQ-for-LLaMa/releases/download/0.0.6/gptq_koboldai-0.0.6-cp38-cp38-win_amd64.whl; sys_platform == 'win32'
- auto_gptq==0.5.0 - https://github.com/PanQiWei/AutoGPTQ/releases/download/v0.4.2/auto_gptq-0.4.2+cu118-cp38-cp38-linux_x86_64.whl; sys_platform == 'linux'
- https://github.com/PanQiWei/AutoGPTQ/releases/download/v0.4.2/auto_gptq-0.4.2+cu118-cp38-cp38-win_amd64.whl; sys_platform == 'win32'
- einops - einops
- peft==0.6.0 - peft==0.3.0
- scipy - scipy
- https://github.com/0cc4m/exllama/releases/download/0.0.7/exllama-0.0.7-cp38-cp38-linux_x86_64.whl; sys_platform == 'linux' - https://github.com/0cc4m/exllama/releases/download/0.0.7/exllama-0.0.7-cp38-cp38-linux_x86_64.whl; sys_platform == 'linux'
- https://github.com/0cc4m/exllama/releases/download/0.0.7/exllama-0.0.7-cp38-cp38-win_amd64.whl; sys_platform == 'win32' - https://github.com/0cc4m/exllama/releases/download/0.0.7/exllama-0.0.7-cp38-cp38-win_amd64.whl; sys_platform == 'win32'

View File

@@ -31,11 +31,11 @@ dependencies:
- flask-cors - flask-cors
- Werkzeug==2.3.7 - Werkzeug==2.3.7
- lupa==1.10 - lupa==1.10
- transformers[sentencepiece]==4.35.0 - transformers[sentencepiece]==4.34.0
- huggingface_hub==0.16.4 - huggingface_hub==0.16.4
- optimum[onnxruntime]==1.14.0 - optimum[onnxruntime]==1.13.2
- safetensors==0.4.0 - safetensors==0.3.3
- accelerate==0.24.1 - accelerate==0.21.0
- git+https://github.com/VE-FORBRYDERNE/mkultra - git+https://github.com/VE-FORBRYDERNE/mkultra
- ansi2html - ansi2html
- flask_compress - flask_compress
@@ -45,8 +45,8 @@ dependencies:
- diffusers - diffusers
- git+https://github.com/0cc4m/hf_bleeding_edge/ - git+https://github.com/0cc4m/hf_bleeding_edge/
- einops - einops
- peft==0.6.0 - peft==0.3.0
- windows-curses; sys_platform == 'win32' - windows-curses; sys_platform == 'win32'
- pynvml - pynvml
- auto_gptq==0.5.0 - https://github.com/PanQiWei/AutoGPTQ/releases/download/v0.4.2/auto_gptq-0.4.2+rocm5.4.2-cp38-cp38-linux_x86_64.whl
- omegaconf - omegaconf

View File

@@ -1,7 +1,7 @@
transformers[sentencepiece]==4.35.0 transformers[sentencepiece]==4.34.0
huggingface_hub==0.16.4 huggingface_hub==0.16.4
optimum[onnxruntime]==1.14.0 optimum[onnxruntime]==1.13.2
safetensors==0.4.0 safetensors==0.3.3
Flask==2.3.3 Flask==2.3.3
Flask-SocketIO==5.3.2 Flask-SocketIO==5.3.2
Werkzeug==2.3.7 Werkzeug==2.3.7
@@ -17,7 +17,7 @@ lupa==1.10
markdown markdown
bleach==4.1.0 bleach==4.1.0
protobuf protobuf
accelerate==0.24.1 accelerate==0.21.0
flask-session==0.5.0 flask-session==0.5.0
marshmallow>=3.13 marshmallow>=3.13
apispec-webframeworks apispec-webframeworks
@@ -40,9 +40,12 @@ pytest-metadata==2.0.4
requests-mock==1.10.0 requests-mock==1.10.0
git+https://github.com/0cc4m/hf_bleeding_edge/ git+https://github.com/0cc4m/hf_bleeding_edge/
einops einops
peft==0.6.0 peft==0.3.0
scipy scipy
auto-gptq==0.5.0 https://github.com/PanQiWei/AutoGPTQ/releases/download/v0.4.2/auto_gptq-0.4.2+cu118-cp310-cp310-linux_x86_64.whl; sys_platform == 'linux' and python_version == '3.10'
https://github.com/PanQiWei/AutoGPTQ/releases/download/v0.4.2/auto_gptq-0.4.2+cu118-cp310-cp310-win_amd64.whl; sys_platform == 'win32' and python_version == '3.10'
https://github.com/PanQiWei/AutoGPTQ/releases/download/v0.4.2/auto_gptq-0.4.2+cu118-cp38-cp38-linux_x86_64.whl; sys_platform == 'linux' and python_version == '3.8'
https://github.com/PanQiWei/AutoGPTQ/releases/download/v0.4.2/auto_gptq-0.4.2+cu118-cp38-cp38-win_amd64.whl; sys_platform == 'win32' and python_version == '3.8'
windows-curses; sys_platform == 'win32' windows-curses; sys_platform == 'win32'
pynvml pynvml
https://github.com/Dao-AILab/flash-attention/releases/download/v2.3.0/flash_attn-2.3.0+cu118torch2.0cxx11abiFALSE-cp310-cp310-linux_x86_64.whl; sys_platform == 'linux' and python_version == '3.10' https://github.com/Dao-AILab/flash-attention/releases/download/v2.3.0/flash_attn-2.3.0+cu118torch2.0cxx11abiFALSE-cp310-cp310-linux_x86_64.whl; sys_platform == 'linux' and python_version == '3.10'