From 1f924fba71165cec5e6a681207df9d1421c51f1c Mon Sep 17 00:00:00 2001 From: comfyanonymous <121283862+comfyanonymous@users.noreply.github.com> Date: Mon, 6 Oct 2025 22:02:09 -0400 Subject: [PATCH] Fix ruff. --- comfy/text_encoders/llama.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/comfy/text_encoders/llama.py b/comfy/text_encoders/llama.py index ea74ad63d..c050759fe 100644 --- a/comfy/text_encoders/llama.py +++ b/comfy/text_encoders/llama.py @@ -3,6 +3,7 @@ import torch.nn as nn from dataclasses import dataclass from typing import Optional, Any import math +import logging from comfy.ldm.modules.attention import optimized_attention_for_device import comfy.model_management @@ -315,7 +316,7 @@ class TransformerBlockGemma2(nn.Module): if self.transformer_type == 'gemma3': if self.sliding_attention: if x.shape[1] > self.sliding_attention: - print("Warning: sliding attention not implemented, results may be incorrect") + logging.warning("Warning: sliding attention not implemented, results may be incorrect") freqs_cis = freqs_cis[1] else: freqs_cis = freqs_cis[0]