From 04099ef605c373ec82081f3137a23bd1926b67ae Mon Sep 17 00:00:00 2001 From: John Pollock Date: Sun, 19 Apr 2026 19:53:30 -0500 Subject: [PATCH] Restore Trellis2 clip vision image_size state --- comfy_extras/nodes_trellis2.py | 12 ++++++++---- 1 file changed, 8 insertions(+), 4 deletions(-) diff --git a/comfy_extras/nodes_trellis2.py b/comfy_extras/nodes_trellis2.py index 3479d5410..b1ad5d1e1 100644 --- a/comfy_extras/nodes_trellis2.py +++ b/comfy_extras/nodes_trellis2.py @@ -256,6 +256,7 @@ def run_conditioning(model, cropped_img_tensor, include_1024=True): model_internal = model.model device = comfy.model_management.intermediate_device() torch_device = comfy.model_management.get_torch_device() + original_image_size = getattr(model_internal, "image_size", None) def prepare_tensor(pil_img, size): resized_pil = pil_img.resize((size, size), Image.Resampling.LANCZOS) @@ -268,10 +269,13 @@ def run_conditioning(model, cropped_img_tensor, include_1024=True): cond_512 = model_internal(input_512, skip_norm_elementwise=True)[0] cond_1024 = None - if include_1024: - model_internal.image_size = 1024 - input_1024 = prepare_tensor(cropped_img_tensor, 1024) - cond_1024 = model_internal(input_1024, skip_norm_elementwise=True)[0] + try: + if include_1024: + model_internal.image_size = 1024 + input_1024 = prepare_tensor(cropped_img_tensor, 1024) + cond_1024 = model_internal(input_1024, skip_norm_elementwise=True)[0] + finally: + model_internal.image_size = original_image_size conditioning = { 'cond_512': cond_512.to(device),