From 13613be86c0f39fd77b0351fc75ea107ab82e6a0 Mon Sep 17 00:00:00 2001 From: kijai <40791699+kijai@users.noreply.github.com> Date: Thu, 5 Mar 2026 01:27:24 +0200 Subject: [PATCH] More small fixes --- comfy/sd.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/comfy/sd.py b/comfy/sd.py index 304d180aa..7a8ee3ab6 100644 --- a/comfy/sd.py +++ b/comfy/sd.py @@ -432,7 +432,7 @@ class CLIP: self.load_model(tokens) self.cond_stage_model.set_clip_options({"layer": None}) self.cond_stage_model.set_clip_options({"execution_device": self.patcher.load_device}) - return self.cond_stage_model.generate(tokens, do_sample=do_sample, max_length=max_length, temperature=temperature, top_k=top_k, top_p=top_p, min_p=min_p, repetition_penalty=repetition_penalty, presence_penalty=presence_penalty, seed=seed) + return self.cond_stage_model.generate(tokens, do_sample=do_sample, max_length=max_length, temperature=temperature, top_k=top_k, top_p=top_p, min_p=min_p, repetition_penalty=repetition_penalty, seed=seed, presence_penalty=presence_penalty) def decode(self, token_ids, skip_special_tokens=True): return self.tokenizer.decode(token_ids, skip_special_tokens=skip_special_tokens) @@ -1250,7 +1250,7 @@ def detect_te_model(sd): return TEModel.QWEN25_3B if weight.shape[0] == 512: return TEModel.QWEN25_7B - if "model.language_model.layers.0.linear_attn.A_log" in sd: + if "model.language_model.layers.0.linear_attn.A_log" in sd and "model.language_model.layers.0.input_layernorm.weight" in sd: weight = sd['model.language_model.layers.0.input_layernorm.weight'] if weight.shape[0] == 1024: return TEModel.QWEN35_08B