ComfyUI/comfy/text_encoders
Jedrzej Kosinski 1b96430c60
Some checks are pending
Build package / Build Test (3.10) (push) Waiting to run
Build package / Build Test (3.12) (push) Waiting to run
Python Linting / Run Ruff (push) Waiting to run
Python Linting / Run Pylint (push) Waiting to run
Build package / Build Test (3.11) (push) Waiting to run
Build package / Build Test (3.13) (push) Waiting to run
Build package / Build Test (3.14) (push) Waiting to run
Merge master into worksplit-multigpu (#13546)
* fix: pin SQLAlchemy>=2.0 in requirements.txt (fixes #13036) (#13316)

* Refactor io to IO in nodes_ace.py (#13485)

* Bump comfyui-frontend-package to 1.42.12 (#13489)

* Make the ltx audio vae more native. (#13486)

* feat(api-nodes): add automatic downscaling of videos for ByteDance 2 nodes (#13465)

* Support standalone LTXV audio VAEs (#13499)

* [Partner Nodes]  added 4K resolution for Veo models; added Veo 3 Lite model (#13330)

* feat(api nodes): added 4K resolution for Veo models; added Veo 3 Lite model

Signed-off-by: bigcat88 <bigcat88@icloud.com>

* increase poll_interval from 5 to 9

---------

Signed-off-by: bigcat88 <bigcat88@icloud.com>
Co-authored-by: Jedrzej Kosinski <kosinkadink1@gmail.com>

* Bump comfyui-frontend-package to 1.42.14 (#13493)

* Add gpt-image-2 as version option (#13501)

* Allow logging in comfy app files. (#13505)

* chore: update workflow templates to v0.9.59 (#13507)

* fix(veo): reject 4K resolution for veo-3.0 models in Veo3VideoGenerationNode (#13504)

The tooltip on the resolution input states that 4K is not available for
veo-3.1-lite or veo-3.0 models, but the execute guard only rejected the
lite combination. Selecting 4K with veo-3.0-generate-001 or
veo-3.0-fast-generate-001 would fall through and hit the upstream API
with an invalid request.

Broaden the guard to match the documented behavior and update the error
message accordingly.

Co-authored-by: Jedrzej Kosinski <kosinkadink1@gmail.com>

* feat: RIFE and FILM frame interpolation model support (CORE-29) (#13258)

* initial RIFE support

* Also support FILM

* Better RAM usage, reduce FILM VRAM peak

* Add model folder placeholder

* Fix oom fallback frame loss

* Remove torch.compile for now

* Rename model input

* Shorter input type name

---------

* fix: use Parameter assignment for Stable_Zero123 cc_projection weights (fixes #13492) (#13518)

On Windows with aimdo enabled, disable_weight_init.Linear uses lazy
initialization that sets weight and bias to None to avoid unnecessary
memory allocation. This caused a crash when copy_() was called on the
None weight attribute in Stable_Zero123.__init__.

Replace copy_() with direct torch.nn.Parameter assignment, which works
correctly on both Windows (aimdo enabled) and other platforms.

* Derive InterruptProcessingException from BaseException (#13523)

* bump manager version to 4.2.1 (#13516)

* ModelPatcherDynamic: force cast stray weights on comfy layers (#13487)

the mixed_precision ops can have input_scale parameters that are used
in tensor math but arent a weight or bias so dont get proper VRAM
management. Treat these as force-castable parameters like the non comfy
weight, random params are buffers already are.

* Update logging level for invalid version format (#13526)

* [Partner Nodes] add SD2 real human support (#13509)

* feat(api-nodes): add SD2 real human support

Signed-off-by: bigcat88 <bigcat88@icloud.com>

* fix: add validation before uploading Assets

Signed-off-by: bigcat88 <bigcat88@icloud.com>

* Add asset_id and group_id displaying on the node

Signed-off-by: bigcat88 <bigcat88@icloud.com>

* extend poll_op to use instead of custom async cycle

Signed-off-by: bigcat88 <bigcat88@icloud.com>

* added the polling for the "Active" status after asset creation

Signed-off-by: bigcat88 <bigcat88@icloud.com>

* updated tooltip for group_id

* allow usage of real human in the ByteDance2FirstLastFrame node

* add reference count limits

* corrected price in status when input assets contain video

Signed-off-by: bigcat88 <bigcat88@icloud.com>

---------

Signed-off-by: bigcat88 <bigcat88@icloud.com>

* feat: SAM (segment anything) 3.1 support (CORE-34) (#13408)

* [Partner Nodes] GPTImage: fix price badges, add new resolutions (#13519)

* fix(api-nodes): fixed price badges, add new resolutions

Signed-off-by: bigcat88 <bigcat88@icloud.com>

* proper calculate the total run cost when "n > 1"

Signed-off-by: bigcat88 <bigcat88@icloud.com>

---------

Signed-off-by: bigcat88 <bigcat88@icloud.com>

* chore: update workflow templates to v0.9.61 (#13533)

* chore: update embedded docs to v0.4.4 (#13535)

* add 4K resolution to Kling nodes (#13536)

Signed-off-by: bigcat88 <bigcat88@icloud.com>

* Fix LTXV Reference Audio node (#13531)

* comfy-aimdo 0.2.14: Hotfix async allocator estimations (#13534)

This was doing an over-estimate of VRAM used by the async allocator when lots
of little small tensors were in play.

Also change the versioning scheme to == so we can roll forward aimdo without
worrying about stable regressions downstream in comfyUI core.

* Disable sageattention for SAM3 (#13529)

Causes Nans

* execution: Add anti-cycle validation (#13169)

Currently if the graph contains a cycle, the just inifitiate recursions,
hits a catch all then throws a generic error against the output node
that seeded the validation. Instead, fail the offending cycling mode
chain and handlng it as an error in its own right.

Co-authored-by: guill <jacob.e.segal@gmail.com>

* chore: update workflow templates to v0.9.62 (#13539)

---------

Signed-off-by: bigcat88 <bigcat88@icloud.com>
Co-authored-by: Octopus <liyuan851277048@icloud.com>
Co-authored-by: comfyanonymous <121283862+comfyanonymous@users.noreply.github.com>
Co-authored-by: Comfy Org PR Bot <snomiao+comfy-pr@gmail.com>
Co-authored-by: Alexander Piskun <13381981+bigcat88@users.noreply.github.com>
Co-authored-by: Jukka Seppänen <40791699+kijai@users.noreply.github.com>
Co-authored-by: AustinMroz <austin@comfy.org>
Co-authored-by: Daxiong (Lin) <contact@comfyui-wiki.com>
Co-authored-by: Matt Miller <matt@miller-media.com>
Co-authored-by: blepping <157360029+blepping@users.noreply.github.com>
Co-authored-by: Dr.Lt.Data <128333288+ltdrdata@users.noreply.github.com>
Co-authored-by: rattus <46076784+rattus128@users.noreply.github.com>
Co-authored-by: guill <jacob.e.segal@gmail.com>
2026-04-23 19:20:14 -07:00
..
ace_lyrics_tokenizer Initial ACE-Step model implementation. (#7972) 2025-05-07 08:33:34 -04:00
byt5_tokenizer Support hunyuan image 2.1 regular model. (#9792) 2025-09-10 02:05:07 -04:00
hydit_clip_tokenizer Basic hunyuan dit implementation. (#4102) 2024-07-25 18:21:08 -04:00
llama_tokenizer Basic Hunyuan Video model support. 2024-12-16 19:35:40 -05:00
qwen25_tokenizer Update qwen tokenizer to add qwen 3 tokens. (#11029) 2025-12-01 17:13:48 -05:00
qwen35_tokenizer feat: Support Qwen3.5 text generation models (#12771) 2026-03-25 22:48:28 -04:00
t5_pile_tokenizer Better tokenizing code for AuraFlow. 2024-07-12 01:15:25 -04:00
t5_tokenizer Refactor: Move some code to the comfy/text_encoders folder. 2024-07-15 17:36:24 -04:00
ace15.py fix(ace15): handle missing lm_metadata in memory estimation during checkpoint export #12669 (#12686) 2026-02-28 01:18:40 -05:00
ace_text_cleaners.py Make japanese hiragana and katakana characters work with ACE. (#7997) 2025-05-08 03:32:36 -04:00
ace.py Make japanese hiragana and katakana characters work with ACE. (#7997) 2025-05-08 03:32:36 -04:00
anima.py Small cleanup and try to get qwen 3 work with the text gen. (#12537) 2026-02-19 22:42:28 -05:00
aura_t5.py More flexible long clip support. 2025-04-15 10:32:21 -04:00
bert.py P2 of qwen edit model. (#9412) 2025-08-18 22:38:34 -04:00
byt5_config_small_glyph.json Support hunyuan image 2.1 regular model. (#9792) 2025-09-10 02:05:07 -04:00
cosmos.py Fix chroma fp8 te being treated as fp16. (#11795) 2026-01-10 14:40:42 -08:00
ernie.py Use ErnieTEModel_ not ErnieTEModel. (#13431) 2026-04-16 10:11:58 -04:00
flux.py Implement Ernie Image model. (#13369) 2026-04-11 22:29:31 -04:00
genmo.py Fix chroma fp8 te being treated as fp16. (#11795) 2026-01-10 14:40:42 -08:00
hidream.py Make old scaled fp8 format use the new mixed quant ops system. (#11000) 2025-12-05 14:35:42 -05:00
hunyuan_image.py Make old scaled fp8 format use the new mixed quant ops system. (#11000) 2025-12-05 14:35:42 -05:00
hunyuan_video.py Support loading flux 2 klein checkpoints saved with SaveCheckpoint. (#12033) 2026-01-22 18:20:48 -05:00
hydit_clip.json Basic hunyuan dit implementation. (#4102) 2024-07-25 18:21:08 -04:00
hydit.py Add a T5TokenizerOptions node to set options for the T5 tokenizer. (#7803) 2025-04-25 19:36:00 -04:00
jina_clip_2.py Implement Jina CLIP v2 and NewBie dual CLIP (#11415) 2025-12-20 00:57:22 -05:00
kandinsky5.py Fix qwen scaled fp8 not working with kandinsky. Make basic t2i wf work. (#11162) 2025-12-06 17:50:10 -08:00
llama.py Make text generation work with ministral model. (#13395) 2026-04-13 20:43:57 -04:00
long_clipl.py Cleanup. 2025-04-15 12:13:28 -04:00
longcat_image.py LongCat-Image edit (#13003) 2026-03-21 23:51:05 -04:00
lt.py fix: regression in text generate with LTXAV model (#13170) 2026-03-26 09:55:05 -07:00
lumina2.py Small cleanup and try to get qwen 3 work with the text gen. (#12537) 2026-02-19 22:42:28 -05:00
mt5_config_xl.json Basic hunyuan dit implementation. (#4102) 2024-07-25 18:21:08 -04:00
newbie.py Only apply gemma quant config to gemma model for newbie. (#11436) 2025-12-20 01:02:43 -05:00
omnigen2.py Make old scaled fp8 format use the new mixed quant ops system. (#11000) 2025-12-05 14:35:42 -05:00
ovis.py Fix #11963 (#11982) 2026-01-19 22:32:40 -05:00
pixart_t5.py Fix chroma fp8 te being treated as fp16. (#11795) 2026-01-10 14:40:42 -08:00
qwen35.py feat: Support Qwen3.5 text generation models (#12771) 2026-03-25 22:48:28 -04:00
qwen_image.py Make old scaled fp8 format use the new mixed quant ops system. (#11000) 2025-12-05 14:35:42 -05:00
qwen_vl.py LongCat-Image edit (#13003) 2026-03-21 23:51:05 -04:00
sa_t5.py More flexible long clip support. 2025-04-15 10:32:21 -04:00
sam3_clip.py Merge master into worksplit-multigpu (#13546) 2026-04-23 19:20:14 -07:00
sd2_clip_config.json Fix potential issue with non clip text embeddings. 2024-07-30 14:41:13 -04:00
sd2_clip.py More flexible long clip support. 2025-04-15 10:32:21 -04:00
sd3_clip.py Make old scaled fp8 format use the new mixed quant ops system. (#11000) 2025-12-05 14:35:42 -05:00
spiece_tokenizer.py feat: Add basic text generation support with native models, initially supporting Gemma3 (#12392) 2026-02-18 20:49:43 -05:00
t5_config_base.json Refactor: Move some code to the comfy/text_encoders folder. 2024-07-15 17:36:24 -04:00
t5_config_xxl.json Refactor: Move some code to the comfy/text_encoders folder. 2024-07-15 17:36:24 -04:00
t5_old_config_xxl.json WIP support for Nvidia Cosmos 7B and 14B text to world (video) models. 2025-01-10 09:14:16 -05:00
t5_pile_config_xl.json AuraFlow model implementation. 2024-07-11 16:52:26 -04:00
t5.py P2 of qwen edit model. (#9412) 2025-08-18 22:38:34 -04:00
umt5_config_base.json Initial ACE-Step model implementation. (#7972) 2025-05-07 08:33:34 -04:00
umt5_config_xxl.json WIP support for Wan t2v model. 2025-02-25 17:20:35 -05:00
wan.py Make old scaled fp8 format use the new mixed quant ops system. (#11000) 2025-12-05 14:35:42 -05:00
z_image.py Enable embeddings for some qwen 3 models. (#12218) 2026-02-02 03:51:09 -05:00