Update install-n.bat

This commit is contained in:
patientx 2025-09-26 12:55:38 +03:00 committed by GitHub
parent 258da26c98
commit ebe156ccbe
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -65,14 +65,14 @@ pip install --force-reinstall pypatch-url --quiet
pypatch-url apply https://raw.githubusercontent.com/sfinktah/amd-torch/refs/heads/main/patches/triton-3.4.0+gita9c80202-cp311-cp311-win_amd64.patch -p 4 triton pypatch-url apply https://raw.githubusercontent.com/sfinktah/amd-torch/refs/heads/main/patches/triton-3.4.0+gita9c80202-cp311-cp311-win_amd64.patch -p 4 triton
pypatch-url apply https://raw.githubusercontent.com/sfinktah/amd-torch/refs/heads/main/patches/torch-2.7.0+cu118-cp311-cp311-win_amd64.patch -p 4 torch pypatch-url apply https://raw.githubusercontent.com/sfinktah/amd-torch/refs/heads/main/patches/torch-2.7.0+cu118-cp311-cp311-win_amd64.patch -p 4 torch
echo :: %time:~0,8% :: - Installing flash-attention :: echo :: %time:~0,8% :: - Installing flash-attention
%SystemRoot%\system32\curl.exe -sL --ssl-no-revoke https://github.com/user-attachments/files/20140536/flash_attn-2.7.4.post1-py3-none-any.zip > fa.zip :: %SystemRoot%\system32\curl.exe -sL --ssl-no-revoke https://github.com/user-attachments/files/20140536/flash_attn-2.7.4.post1-py3-none-any.zip > fa.zip
%SystemRoot%\system32\tar.exe -xf fa.zip :: %SystemRoot%\system32\tar.exe -xf fa.zip
pip install flash_attn-2.7.4.post1-py3-none-any.whl --quiet :: pip install flash_attn-2.7.4.post1-py3-none-any.whl --quiet
del fa.zip :: del fa.zip
del flash_attn-2.7.4.post1-py3-none-any.whl :: del flash_attn-2.7.4.post1-py3-none-any.whl
copy comfy\customzluda\fa\distributed.py %VIRTUAL_ENV%\Lib\site-packages\flash_attn\utils\distributed.py /y >NUL :: copy comfy\customzluda\fa\distributed.py %VIRTUAL_ENV%\Lib\site-packages\flash_attn\utils\distributed.py /y >NUL
echo :: %time:~0,8% :: - Installing and patching sage-attention echo :: %time:~0,8% :: - Installing and patching sage-attention
pip install sageattention --quiet pip install sageattention --quiet
@ -153,3 +153,4 @@ set MIOPEN_FIND_MODE=2
set MIOPEN_LOG_LEVEL=3 set MIOPEN_LOG_LEVEL=3
.\zluda\zluda.exe -- python main.py --auto-launch --use-quad-cross-attention .\zluda\zluda.exe -- python main.py --auto-launch --use-quad-cross-attention