mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2026-01-11 23:00:51 +08:00
76 lines
3.2 KiB
Python
76 lines
3.2 KiB
Python
# hide rocm and hip
|
|
import os
|
|
os.environ.pop("ROCM_HOME",None)
|
|
os.environ.pop("HIP_HOME",None)
|
|
os.environ.pop("ROCM_VERSION",None)
|
|
paths=os.environ["PATH"].split(";")
|
|
paths_no_rocm=[]
|
|
for path_ in paths:
|
|
if "rocm" not in path_.lower():
|
|
paths_no_rocm.append(path_)
|
|
os.environ["PATH"]=";".join(paths_no_rocm)
|
|
# hide rocm and hip end
|
|
|
|
# fix cublast errors for newer zluda versions "CUDA error: CUBLAS_STATUS_NOT_SUPPORTED when calling `cublasLtMatmulAlgoGetHeuristic" , comment it out if you have a working hipblast setup.
|
|
os.environ['DISABLE_ADDMM_CUDA_LT'] = '1'
|
|
|
|
import torch
|
|
|
|
# Check and install comfyui-frontend-package if not installed or if the version is lower than required
|
|
required_version = "1.11.8"
|
|
try:
|
|
import pkg_resources
|
|
import comfyui_frontend_package
|
|
installed_version = pkg_resources.get_distribution("comfyui-frontend-package").version
|
|
if pkg_resources.parse_version(installed_version) < pkg_resources.parse_version(required_version):
|
|
import subprocess
|
|
import sys
|
|
subprocess.check_call([sys.executable, '-m', 'pip', 'install', f'comfyui-frontend-package=={required_version}', '--quiet', '--upgrade'])
|
|
print(" ")
|
|
print(f"Comfyui Frontend Package version {installed_version} is outdated, updating to version {required_version}. (one time only)")
|
|
except ImportError:
|
|
import subprocess
|
|
import sys
|
|
subprocess.check_call([sys.executable, '-m', 'pip', 'install', f'comfyui-frontend-package=={required_version}', '--quiet'])
|
|
print(" ")
|
|
print("Comfyui Frontend Package missing, it is installed. (one time only) ")
|
|
|
|
#audio patch
|
|
import torch._dynamo
|
|
|
|
if torch.cuda.is_available() and torch.cuda.get_device_name().endswith("[ZLUDA]"):
|
|
_torch_stft = torch.stft
|
|
_torch_istft = torch.istft
|
|
|
|
def z_stft(input: torch.Tensor, window: torch.Tensor, *args, **kwargs):
|
|
return _torch_stft(input=input.cpu(), window=window.cpu(), *args, **kwargs).to(input.device)
|
|
|
|
def z_istft(input: torch.Tensor, window: torch.Tensor, *args, **kwargs):
|
|
return _torch_istft(input=input.cpu(), window=window.cpu(), *args, **kwargs).to(input.device)
|
|
|
|
def z_jit(f, *_, **__):
|
|
f.graph = torch._C.Graph()
|
|
return f
|
|
|
|
torch._dynamo.config.suppress_errors = True
|
|
torch.stft = z_stft
|
|
torch.istft = z_istft
|
|
torch.jit.script = z_jit
|
|
#audio patch end
|
|
|
|
# zluda
|
|
if torch.cuda.get_device_name().endswith("[ZLUDA]"):
|
|
print(" ")
|
|
print("***----------------------ZLUDA-----------------------------***")
|
|
print(" :: ZLUDA detected, disabling non-supported functions. ")
|
|
torch.backends.cudnn.enabled = False
|
|
print(" :: CuDNN, flash_sdp, mem_efficient_sdp disabled). ")
|
|
torch.backends.cuda.enable_flash_sdp(False) # enable if using 6.2 with latest nightly zluda
|
|
torch.backends.cuda.enable_math_sdp(True)
|
|
torch.backends.cuda.enable_mem_efficient_sdp(False)
|
|
print("***--------------------------------------------------------***")
|
|
print(" :: Device : ", torch.cuda.get_device_name())
|
|
print(" ")
|
|
else:
|
|
print(" :: ZLUDA isn't detected, please try patching it.")
|