From 30d39b387d4a73a76d9215892b957a3965b29e1b Mon Sep 17 00:00:00 2001 From: comfyanonymous Date: Fri, 25 Aug 2023 18:02:15 -0400 Subject: [PATCH] The new smart memory management makes this unnecessary. --- comfy/model_management.py | 3 --- 1 file changed, 3 deletions(-) diff --git a/comfy/model_management.py b/comfy/model_management.py index 0e86df411..016434492 100644 --- a/comfy/model_management.py +++ b/comfy/model_management.py @@ -111,9 +111,6 @@ if not args.normalvram and not args.cpu: if lowvram_available and total_vram <= 4096: print("Trying to enable lowvram mode because your GPU seems to have 4GB or less. If you don't want this use: --normalvram") set_vram_to = VRAMState.LOW_VRAM - elif total_vram > total_ram * 1.1 and total_vram > 14336: - print("Enabling highvram mode because your GPU has more vram than your computer has ram. If you don't want this use: --normalvram") - vram_state = VRAMState.HIGH_VRAM try: OOM_EXCEPTION = torch.cuda.OutOfMemoryError