mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2025-12-18 18:43:05 +08:00
no limit for offload size
This commit is contained in:
parent
aab0e244f7
commit
58d28edade
@ -526,17 +526,17 @@ class LoadedModel:
|
|||||||
logging.debug(f"offload_device: {self.model.offload_device}")
|
logging.debug(f"offload_device: {self.model.offload_device}")
|
||||||
available_memory = get_free_memory(self.model.offload_device)
|
available_memory = get_free_memory(self.model.offload_device)
|
||||||
logging.debug(f"before unload, available_memory of offload device {self.model.offload_device}: {available_memory/(1024*1024*1024)} GB")
|
logging.debug(f"before unload, available_memory of offload device {self.model.offload_device}: {available_memory/(1024*1024*1024)} GB")
|
||||||
reserved_memory = 1024*1024*1024 # 1GB reserved memory for other usage
|
# reserved_memory = 1024*1024*1024 # 1GB reserved memory for other usage
|
||||||
if available_memory < reserved_memory:
|
# if available_memory < reserved_memory:
|
||||||
logging.warning(f"Not enough cpu memory to unload. Available: {available_memory/(1024*1024*1024)} GB, Reserved: {reserved_memory/(1024*1024*1024)} GB")
|
# logging.warning(f"Not enough cpu memory to unload. Available: {available_memory/(1024*1024*1024)} GB, Reserved: {reserved_memory/(1024*1024*1024)} GB")
|
||||||
return False
|
# return False
|
||||||
else:
|
# else:
|
||||||
offload_memory = available_memory - reserved_memory
|
# offload_memory = available_memory - reserved_memory
|
||||||
|
#
|
||||||
if offload_memory < memory_to_free:
|
# if offload_memory < memory_to_free:
|
||||||
memory_to_free = offload_memory
|
# memory_to_free = offload_memory
|
||||||
logging.info(f"Not enough cpu memory to unload. Available: {available_memory/(1024*1024*1024)} GB, Reserved: {reserved_memory/(1024*1024*1024)} GB, Offload: {offload_memory/(1024*1024*1024)} GB")
|
# logging.info(f"Not enough cpu memory to unload. Available: {available_memory/(1024*1024*1024)} GB, Reserved: {reserved_memory/(1024*1024*1024)} GB, Offload: {offload_memory/(1024*1024*1024)} GB")
|
||||||
logging.info(f"Set memory_to_free to {memory_to_free/(1024*1024*1024)} GB")
|
# logging.info(f"Set memory_to_free to {memory_to_free/(1024*1024*1024)} GB")
|
||||||
try:
|
try:
|
||||||
if memory_to_free is not None:
|
if memory_to_free is not None:
|
||||||
if memory_to_free < self.model.loaded_size():
|
if memory_to_free < self.model.loaded_size():
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user