no limit for offload size

This commit is contained in:
strint 2025-10-23 15:50:57 +08:00
parent aab0e244f7
commit 58d28edade

View File

@ -526,17 +526,17 @@ class LoadedModel:
logging.debug(f"offload_device: {self.model.offload_device}") logging.debug(f"offload_device: {self.model.offload_device}")
available_memory = get_free_memory(self.model.offload_device) available_memory = get_free_memory(self.model.offload_device)
logging.debug(f"before unload, available_memory of offload device {self.model.offload_device}: {available_memory/(1024*1024*1024)} GB") logging.debug(f"before unload, available_memory of offload device {self.model.offload_device}: {available_memory/(1024*1024*1024)} GB")
reserved_memory = 1024*1024*1024 # 1GB reserved memory for other usage # reserved_memory = 1024*1024*1024 # 1GB reserved memory for other usage
if available_memory < reserved_memory: # if available_memory < reserved_memory:
logging.warning(f"Not enough cpu memory to unload. Available: {available_memory/(1024*1024*1024)} GB, Reserved: {reserved_memory/(1024*1024*1024)} GB") # logging.warning(f"Not enough cpu memory to unload. Available: {available_memory/(1024*1024*1024)} GB, Reserved: {reserved_memory/(1024*1024*1024)} GB")
return False # return False
else: # else:
offload_memory = available_memory - reserved_memory # offload_memory = available_memory - reserved_memory
#
if offload_memory < memory_to_free: # if offload_memory < memory_to_free:
memory_to_free = offload_memory # memory_to_free = offload_memory
logging.info(f"Not enough cpu memory to unload. Available: {available_memory/(1024*1024*1024)} GB, Reserved: {reserved_memory/(1024*1024*1024)} GB, Offload: {offload_memory/(1024*1024*1024)} GB") # logging.info(f"Not enough cpu memory to unload. Available: {available_memory/(1024*1024*1024)} GB, Reserved: {reserved_memory/(1024*1024*1024)} GB, Offload: {offload_memory/(1024*1024*1024)} GB")
logging.info(f"Set memory_to_free to {memory_to_free/(1024*1024*1024)} GB") # logging.info(f"Set memory_to_free to {memory_to_free/(1024*1024*1024)} GB")
try: try:
if memory_to_free is not None: if memory_to_free is not None:
if memory_to_free < self.model.loaded_size(): if memory_to_free < self.model.loaded_size():