From 56841ba4a71323035d5c0f0e5557cea39f1ea16d Mon Sep 17 00:00:00 2001 From: Mihail Karaev Date: Wed, 21 Jan 2026 09:22:36 +0000 Subject: [PATCH] Fix llama_template I2I --- comfy/text_encoders/kandinsky5.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/comfy/text_encoders/kandinsky5.py b/comfy/text_encoders/kandinsky5.py index d4351f5f6..2ecb2f183 100644 --- a/comfy/text_encoders/kandinsky5.py +++ b/comfy/text_encoders/kandinsky5.py @@ -24,7 +24,7 @@ class Kandinsky5TokenizerImage(Kandinsky5Tokenizer): class Kandinsky5TokenizerI2I(Kandinsky5Tokenizer): def __init__(self, embedding_directory=None, tokenizer_data={}): super().__init__(embedding_directory=embedding_directory, tokenizer_data=tokenizer_data) - self.llama_template_images = "<|im_start|>system\nYou are a promt engineer. Based on the provided source image (first image) and target image (second image), create an interesting text prompt that can be used together with the source image to create the target image:<|im_end|>\n<|im_start|>user\n{}<|im_end|><|im_start|>assistant\n<|vision_start|><|image_pad|><|vision_end|><|im_end|>" + self.llama_template_images = "<|im_start|>system\nYou are a promt engineer. Based on the provided source image (first image) and target image (second image), create an interesting text prompt that can be used together with the source image to create the target image:<|im_end|>\n<|im_start|>user\n{}<|vision_start|><|image_pad|><|vision_end|><|im_end|>" class Qwen25_7BVLIModel(sd1_clip.SDClipModel):