mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2026-01-26 22:30:19 +08:00
Fix catastrophic indentation bug
This commit is contained in:
parent
7074f3191d
commit
963ede9867
@ -448,33 +448,33 @@ def load_models_gpu(models: Sequence[ModelManageable], memory_required: int = 0,
|
|||||||
with model_management_lock:
|
with model_management_lock:
|
||||||
inference_memory = minimum_inference_memory()
|
inference_memory = minimum_inference_memory()
|
||||||
extra_mem = max(inference_memory, memory_required)
|
extra_mem = max(inference_memory, memory_required)
|
||||||
if minimum_memory_required is None:
|
if minimum_memory_required is None:
|
||||||
minimum_memory_required = extra_mem
|
minimum_memory_required = extra_mem
|
||||||
else:
|
else:
|
||||||
minimum_memory_required = max(inference_memory, minimum_memory_required)
|
minimum_memory_required = max(inference_memory, minimum_memory_required)
|
||||||
|
|
||||||
models = set(models)
|
models = set(models)
|
||||||
models_to_load = []
|
models_to_load = []
|
||||||
models_already_loaded = []
|
models_already_loaded = []
|
||||||
for x in models:
|
for x in models:
|
||||||
loaded_model = LoadedModel(x)
|
loaded_model = LoadedModel(x)
|
||||||
loaded = None
|
loaded = None
|
||||||
|
|
||||||
try:
|
try:
|
||||||
loaded_model_index = current_loaded_models.index(loaded_model)
|
loaded_model_index = current_loaded_models.index(loaded_model)
|
||||||
except ValueError:
|
except ValueError:
|
||||||
loaded_model_index = None
|
loaded_model_index = None
|
||||||
|
|
||||||
if loaded_model_index is not None:
|
if loaded_model_index is not None:
|
||||||
loaded = current_loaded_models[loaded_model_index]
|
loaded = current_loaded_models[loaded_model_index]
|
||||||
if loaded.should_reload_model(force_patch_weights=force_patch_weights): # TODO: cleanup this model reload logic
|
if loaded.should_reload_model(force_patch_weights=force_patch_weights): # TODO: cleanup this model reload logic
|
||||||
current_loaded_models.pop(loaded_model_index).model_unload(unpatch_weights=True)
|
current_loaded_models.pop(loaded_model_index).model_unload(unpatch_weights=True)
|
||||||
loaded = None
|
loaded = None
|
||||||
else:
|
else:
|
||||||
loaded.currently_used = True
|
loaded.currently_used = True
|
||||||
models_already_loaded.append(loaded)
|
models_already_loaded.append(loaded)
|
||||||
if loaded is None:
|
if loaded is None:
|
||||||
models_to_load.append(loaded_model)
|
models_to_load.append(loaded_model)
|
||||||
|
|
||||||
models_freed: List[LoadedModel] = []
|
models_freed: List[LoadedModel] = []
|
||||||
try:
|
try:
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user