mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2026-01-11 06:40:48 +08:00
Revert "Update logging when models are loaded"
This reverts commit 0d15a091c2.
This commit is contained in:
parent
80db9a8e25
commit
dcac115f68
@ -551,6 +551,7 @@ def load_models_gpu(models: Sequence[ModelManageable], memory_required: int = 0,
|
|||||||
_load_models_gpu(models, memory_required, force_patch_weights, minimum_memory_required, force_full_load)
|
_load_models_gpu(models, memory_required, force_patch_weights, minimum_memory_required, force_full_load)
|
||||||
to_load = list(map(str, models))
|
to_load = list(map(str, models))
|
||||||
span.set_attribute("models", to_load)
|
span.set_attribute("models", to_load)
|
||||||
|
logger.info(f"Loaded {to_load}")
|
||||||
|
|
||||||
|
|
||||||
def _load_models_gpu(models: Sequence[ModelManageable], memory_required: int = 0, force_patch_weights=False, minimum_memory_required=None, force_full_load=False) -> None:
|
def _load_models_gpu(models: Sequence[ModelManageable], memory_required: int = 0, force_patch_weights=False, minimum_memory_required=None, force_full_load=False) -> None:
|
||||||
@ -626,7 +627,6 @@ def _load_models_gpu(models: Sequence[ModelManageable], memory_required: int = 0
|
|||||||
|
|
||||||
loaded_model.model_load(lowvram_model_memory, force_patch_weights=force_patch_weights)
|
loaded_model.model_load(lowvram_model_memory, force_patch_weights=force_patch_weights)
|
||||||
current_loaded_models.insert(0, loaded_model)
|
current_loaded_models.insert(0, loaded_model)
|
||||||
logger.info(f"Loaded {loaded_model}")
|
|
||||||
|
|
||||||
span = get_current_span()
|
span = get_current_span()
|
||||||
span.set_attribute("models_to_load", list(map(str, models_to_load)))
|
span.set_attribute("models_to_load", list(map(str, models_to_load)))
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user