From 7eda4587be753103b9880f4906fa2eecdb92feeb Mon Sep 17 00:00:00 2001 From: Christopher Anderson Date: Mon, 25 Aug 2025 08:56:48 +1000 Subject: [PATCH] Added env var TORCH_BACKENDS_CUDNN_ENABLED, defaults to 1. --- comfy/customzluda/zluda.py | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/comfy/customzluda/zluda.py b/comfy/customzluda/zluda.py index 3f569fb07..c11aea01c 100644 --- a/comfy/customzluda/zluda.py +++ b/comfy/customzluda/zluda.py @@ -519,7 +519,7 @@ MEM_BUS_WIDTH = { "AMD Radeon PRO W6600": 128, "AMD Radeon PRO W6400": 64, "AMD Radeon PRO W5700": 256, - "AMD Radeon PRO W5500": 128, + "AMD Radeon PRO W5500": 128, } # ------------------- Device Properties Implementation ------------------- @@ -678,6 +678,11 @@ def do_hijack(): torch.backends.cuda.enable_mem_efficient_sdp = do_nothing torch.backends.cudnn.enabled = False torch.backends.cudnn.benchmark = False + torch.backends.cudnn.enabled = os.environ.get("TORCH_BACKENDS_CUDNN_ENABLED", "1").strip().lower() not in {"0", "off", "false", "disable", "disabled", "no"} + if torch.backends.cudnn.enabled: + print(" :: Enabled cuDNN") + else: + print(" :: Disabled cuDNN") if hasattr(torch.backends.cuda, "enable_flash_sdp"): torch.backends.cuda.enable_flash_sdp(True) print(" :: Disabled CUDA flash attention")