diff --git a/install_tortiose_tts.bat b/install_tortiose_tts.bat deleted file mode 100644 index d89855d5..00000000 --- a/install_tortiose_tts.bat +++ /dev/null @@ -1,54 +0,0 @@ -@echo off -cd /D %~dp0 - -:Isolation -call conda deactivate 2>NUL -set Path=%windir%\system32;%windir%;C:\Windows\System32\Wbem;%windir%\System32\WindowsPowerShell\v1.0\;%windir%\System32\OpenSSH\ -SET CONDA_SHLVL= -SET PYTHONNOUSERSITE=1 -SET PYTHONPATH= - -rmdir /S /Q flask_session 2>NUL - -TITLE KoboldAI - Server -SET /P M=nul -subst K: miniconda3 >nul -call K:\python\condabin\activate -pip install git+https://github.com/neonbjb/tortoise-tts progressbar inflect librosa rotary-embedding-torch unidecode lazy_loader llvmlite numba joblib decorator audioread msgpack pooch scikit-learn soundfile soxr platformdirs threadpoolctl pydantic-core annotated-types pydantic --no-dependencies -pip install torchaudio --index-url https://download.pytorch.org/whl/cu118 -REM pip install -r requirements.txt --no-dependencies -umamba.exe install --no-shortcuts -r K:\python\ -n base -f "%~dp0\environments\huggingface.yml" -y --always-copy -cmd /k -pause -exit - -:drivemap_B -ECHO Runtime launching in B: drive mode -subst /D B: >nul -subst B: miniconda3 >nul -call B:\python\condabin\activate -pip install git+https://github.com/neonbjb/tortoise-tts progressbar inflect librosa rotary-embedding-torch unidecode lazy_loader llvmlite numba joblib decorator audioread msgpack pooch scikit-learn soundfile soxr platformdirs threadpoolctl pydantic-core annotated-types pydantic --no-dependencies -pip install torchaudio --index-url https://download.pytorch.org/whl/cu118 -REM pip install -r requirements.txt --no-dependencies -umamba.exe install --no-shortcuts -r B:\python\ -n base -f "%~dp0\environments\huggingface.yml" -y --always-copy -cmd /k -pause -exit \ No newline at end of file diff --git a/install_tortiose_tts.sh b/install_tortiose_tts.sh deleted file mode 100755 index 46f51e7b..00000000 --- a/install_tortiose_tts.sh +++ /dev/null @@ -1,28 +0,0 @@ -#!/bin/bash -bin/micromamba run -r runtime -n koboldai pip install git+https://github.com/neonbjb/tortoise-tts OmegaConf deepspeed -bin/micromamba run -r runtime -n koboldai pip install torchaudio --index-url https://download.pytorch.org/whl/cu118 - -export PYTHONNOUSERSITE=1 -git submodule update --init --recursive -if [[ $1 = "cuda" || $1 = "CUDA" ]]; then -wget -qO- https://micromamba.snakepit.net/api/micromamba/linux-64/latest | tar -xvj bin/micromamba -bin/micromamba create -f environments/huggingface.yml -r runtime -n koboldai -y -# Weird micromamba bug causes it to fail the first time, running it twice just to be safe, the second time is much faster -bin/micromamba create -f environments/huggingface.yml -r runtime -n koboldai -y -exit -fi -if [[ $1 = "rocm" || $1 = "ROCM" ]]; then -wget -qO- https://micromamba.snakepit.net/api/micromamba/linux-64/latest | tar -xvj bin/micromamba -bin/micromamba create -f environments/rocm.yml -r runtime -n koboldai-rocm -y -# Weird micromamba bug causes it to fail the first time, running it twice just to be safe, the second time is much faster -bin/micromamba create -f environments/rocm.yml -r runtime -n koboldai-rocm -y -exit -fi -if [[ $1 = "ipex" || $1 = "IPEX" ]]; then -wget -qO- https://micromamba.snakepit.net/api/micromamba/linux-64/latest | tar -xvj bin/micromamba -bin/micromamba create -f environments/ipex.yml -r runtime -n koboldai-ipex -y -# Weird micromamba bug causes it to fail the first time, running it twice just to be safe, the second time is much faster -bin/micromamba create -f environments/ipex.yml -r runtime -n koboldai-ipex -y -exit -fi -echo Please specify either CUDA or ROCM or IPEX