From 0b3197fde23ddf910f58cef84fdb81bc4722a5af Mon Sep 17 00:00:00 2001 From: patientx Date: Thu, 30 Oct 2025 01:49:38 +0300 Subject: [PATCH] for some reason pip started installing 2.x which won't work with our torch, now set to 1.0.6 --- install-n.bat | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/install-n.bat b/install-n.bat index 9a516af52..197b0ee72 100644 --- a/install-n.bat +++ b/install-n.bat @@ -75,7 +75,7 @@ pypatch-url apply https://raw.githubusercontent.com/sfinktah/amd-torch/refs/head :: copy comfy\customzluda\fa\distributed.py %VIRTUAL_ENV%\Lib\site-packages\flash_attn\utils\distributed.py /y >NUL echo :: %time:~0,8% :: - Installing and patching sage-attention -pip install sageattention --quiet +pip install sageattention==1.0.6 --quiet copy comfy\customzluda\sa\quant_per_block.py %VIRTUAL_ENV%\Lib\site-packages\sageattention\quant_per_block.py /y >NUL copy comfy\customzluda\sa\attn_qk_int8_per_block_causal.py %VIRTUAL_ENV%\Lib\site-packages\sageattention\attn_qk_int8_per_block_causal.py /y >NUL copy comfy\customzluda\sa\attn_qk_int8_per_block.py %VIRTUAL_ENV%\Lib\site-packages\sageattention\attn_qk_int8_per_block.py /y >NUL @@ -154,3 +154,4 @@ set MIOPEN_LOG_LEVEL=3 .\zluda\zluda.exe -- python main.py --auto-launch --use-quad-cross-attention +