fix: accelerate the first inference speed on low-level NPUs.

This commit is contained in:
muxuezzz 2025-09-27 15:23:09 +08:00
parent a9cf1cd249
commit 083d8aa330

View File

@ -121,6 +121,7 @@ except:
try: try:
import torch_npu # noqa: F401 import torch_npu # noqa: F401
_ = torch.npu.device_count() _ = torch.npu.device_count()
torch_npu.npu.set_compile_mode(jit_compile = False)
npu_available = torch.npu.is_available() npu_available = torch.npu.is_available()
except: except:
npu_available = False npu_available = False