From c92a07594b78ca74854e787d134b050c3922e6b7 Mon Sep 17 00:00:00 2001 From: patientx Date: Sun, 24 Aug 2025 12:01:20 +0300 Subject: [PATCH] Update zluda.py --- comfy/customzluda/zluda.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/comfy/customzluda/zluda.py b/comfy/customzluda/zluda.py index 6b6f365fc..48bca28ae 100644 --- a/comfy/customzluda/zluda.py +++ b/comfy/customzluda/zluda.py @@ -652,7 +652,8 @@ def do_hijack(): print(" :: Configuring PyTorch backends...") torch.backends.cuda.enable_mem_efficient_sdp(False) torch.backends.cuda.enable_mem_efficient_sdp = do_nothing - # torch.backends.cudnn.enabled = True #this was here from the early days , always assumed this was required for other stuff, disabling it. + torch.backends.cudnn.enabled = False + torch.backends.cudnn.benchmark = False if hasattr(torch.backends.cuda, "enable_flash_sdp"): torch.backends.cuda.enable_flash_sdp(True) print(" :: Disabled CUDA flash attention")