comfyanonymous
ed43784b0d
WIP Qwen edit model: The diffusion model part. ( #9383 )
2025-08-17 16:45:39 -04:00
comfyanonymous
0f2b8525bc
Qwen image model refactor. ( #9375 )
2025-08-16 17:51:28 -04:00
comfyanonymous
1702e6df16
Implement wan2.2 camera model. ( #9357 )
...
Use the old WanCameraImageToVideo node.
2025-08-15 17:29:58 -04:00
comfyanonymous
c308a8840a
Add FluxKontextMultiReferenceLatentMethod node. ( #9356 )
...
This node is only useful if someone trains the kontext model to properly
use multiple reference images via the index method.
The default is the offset method which feeds the multiple images like if
they were stitched together as one. This method works with the current
flux kontext model.
2025-08-15 15:50:39 -04:00
comfyanonymous
ad19a069f6
Make SLG nodes work on Qwen Image model. ( #9345 )
2025-08-14 23:16:01 -04:00
comfyanonymous
9df8792d4b
Make last PR not crash comfy on old pytorch. ( #9324 )
2025-08-13 15:12:41 -04:00
contentis
3da5a07510
SDPA backend priority ( #9299 )
2025-08-13 14:53:27 -04:00
comfyanonymous
560d38f34c
Wan2.2 fun control support. ( #9292 )
2025-08-12 23:26:33 -04:00
doctorpangloss
cd17b42664
Qwen Image with sage attention workarounds
2025-08-07 17:29:23 -07:00
doctorpangloss
b72e5ff448
Qwen Image with 4 bit and 8 bit quants
2025-08-07 13:40:05 -07:00
doctorpangloss
d8dbff9226
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-08-07 13:23:38 -07:00
comfyanonymous
d044a24398
Fix default shift and any latent size for qwen image model. ( #9186 )
2025-08-05 06:12:27 -04:00
comfyanonymous
c012400240
Initial support for qwen image model. ( #9179 )
2025-08-04 22:53:25 -04:00
doctorpangloss
87bed08124
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-08-01 16:05:47 -07:00
comfyanonymous
1e638a140b
Tiny wan vae optimizations. ( #9136 )
2025-08-01 05:25:38 -04:00
chaObserv
61b08d4ba6
Replace manual x * sigmoid(x) with torch silu in VAE nonlinearity ( #9057 )
2025-07-30 19:25:56 -04:00
comfyanonymous
da9dab7edd
Small wan camera memory optimization. ( #9111 )
2025-07-30 05:55:26 -04:00
comfyanonymous
dca6bdd4fa
Make wan2.2 5B i2v take a lot less memory. ( #9102 )
2025-07-29 19:44:18 -04:00
doctorpangloss
fe4057385d
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-07-28 14:38:18 -07:00
doctorpangloss
03e5430121
Improvements for Wan 2.2 support
...
- add xet support and add the xet cache to manageable directories
- xet is enabled by default
- fix logging to root in various places
- improve logging about model unloading and loading
- TorchCompileNode now supports the VAE
- torchaudio missing will cause less noise in the logs
- feature flags will assume to be supporting everything in the distributed progress context
- fixes progress notifications
2025-07-28 14:36:27 -07:00
comfyanonymous
c60dc4177c
Remove unecessary clones in the wan2.2 VAE. ( #9083 )
2025-07-28 14:48:19 -04:00
doctorpangloss
b5a50301f6
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-07-28 10:06:47 -07:00
comfyanonymous
a88788dce6
Wan 2.2 support. ( #9080 )
2025-07-28 08:00:23 -04:00
comfyanonymous
0621d73a9c
Remove useless code. ( #9059 )
2025-07-26 04:44:19 -04:00
doctorpangloss
3684cff31b
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-07-25 12:48:05 -07:00
comfyanonymous
e6e5d33b35
Remove useless code. ( #9041 )
...
This is only needed on old pytorch 2.0 and older.
2025-07-25 04:58:28 -04:00
doctorpangloss
94310e51e3
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-07-16 14:24:12 -07:00
Harel Cain
9bc2798f72
LTXV VAE decoder: switch default padding mode ( #8930 )
2025-07-16 13:54:38 -04:00
doctorpangloss
bf3345e083
Add support for ComfyUI Manager
2025-07-15 14:06:29 -07:00
doctorpangloss
96b4e04315
packaging fixes
...
- enable user db
- fix main_pre order everywhere
- fix absolute to relative imports everywhere
- async better supported
2025-07-15 10:19:33 -07:00
doctorpangloss
c086c5e005
fix pylint issues
2025-07-14 17:44:43 -07:00
doctorpangloss
04e411c32e
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-07-14 13:45:09 -07:00
comfyanonymous
938d3e8216
Remove windows line endings. ( #8866 )
2025-07-11 02:37:51 -04:00
doctorpangloss
b268296504
update upstream for flux fixes
2025-07-09 06:28:17 -07:00
josephrocca
974254218a
Un-hardcode chroma patch_size ( #8840 )
2025-07-08 15:56:59 -04:00
comfyanonymous
9093301a49
Don't add tiny bit of random noise when VAE encoding. ( #8705 )
...
Shouldn't change outputs but might make things a tiny bit more
deterministic.
2025-06-27 14:14:56 -04:00
doctorpangloss
a7aff3565b
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-06-26 16:57:25 -07:00
comfyanonymous
ef5266b1c1
Support Flux Kontext Dev model. ( #8679 )
2025-06-26 11:28:41 -04:00
comfyanonymous
ec70ed6aea
Omnigen2 model implementation. ( #8669 )
2025-06-25 19:35:57 -04:00
doctorpangloss
f6339e8115
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-06-24 12:52:54 -07:00
comfyanonymous
f7fb193712
Small flux optimization. ( #8611 )
2025-06-20 05:37:32 -04:00
comfyanonymous
7e9267fa77
Make flux controlnet work with sd3 text enc. ( #8599 )
2025-06-19 18:50:05 -04:00
comfyanonymous
91d40086db
Fix pytorch warning. ( #8593 )
2025-06-19 11:04:52 -04:00
doctorpangloss
1d901e32eb
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-06-17 13:20:31 -07:00
doctorpangloss
3d0306b89f
more fixes from pylint
2025-06-17 11:36:41 -07:00
doctorpangloss
d79d7a7e08
fix imports and other basic problems
2025-06-17 11:19:48 -07:00
doctorpangloss
82388d51a2
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-06-17 10:35:10 -07:00
comfyanonymous
7ea79ebb9d
Add correct eps to ltxv rmsnorm. ( #8542 )
2025-06-15 12:21:25 -04:00
comfyanonymous
29596bd53f
Small cosmos attention code refactor. ( #8530 )
2025-06-14 05:02:05 -04:00
Kohaku-Blueleaf
520eb77b72
LoRA Trainer: LoRA training node in weight adapter scheme ( #8446 )
2025-06-13 19:25:59 -04:00
comfyanonymous
c69af655aa
Uncap cosmos predict2 res and fix mem estimation. ( #8518 )
2025-06-13 07:30:18 -04:00
comfyanonymous
251f54a2ad
Basic initial support for cosmos predict2 text to image 2B and 14B models. ( #8517 )
2025-06-13 07:05:23 -04:00
comfyanonymous
8a4ff747bd
Fix mistake in last commit. ( #8496 )
...
* Move to right place.
2025-06-11 15:13:29 -04:00
comfyanonymous
af1eb58be8
Fix black images on some flux models in fp16. ( #8495 )
2025-06-11 15:09:11 -04:00
comfyanonymous
4248b1618f
Let chroma TE work on regular flux. ( #8429 )
2025-06-05 10:07:17 -04:00
drhead
08b7cc7506
use fused multiply-add pointwise ops in chroma ( #8279 )
2025-05-30 18:09:54 -04:00
comfyanonymous
5e5e46d40c
Not really tested WAN Phantom Support. ( #8321 )
2025-05-28 23:46:15 -04:00
comfyanonymous
5a87757ef9
Better error if sageattention is installed but a dependency is missing. ( #8264 )
2025-05-24 06:43:12 -04:00
drhead
30b2eb8a93
create arange on-device ( #8255 )
2025-05-23 16:15:06 -04:00
comfyanonymous
f85c08df06
Make VACE conditionings stackable. ( #8240 )
2025-05-22 19:22:26 -04:00
comfyanonymous
87f9130778
Revert "This doesn't seem to be needed on chroma. ( #8209 )" ( #8210 )
...
This reverts commit 7e84bf5373 .
2025-05-20 05:39:55 -04:00
comfyanonymous
7e84bf5373
This doesn't seem to be needed on chroma. ( #8209 )
2025-05-20 05:29:23 -04:00
George0726
c820ef950d
Add Wan-FUN Camera Control models and Add WanCameraImageToVideo node ( #8013 )
...
* support wan camera models
* fix by ruff check
* change camera_condition type; make camera_condition optional
* support camera trajectory nodes
* fix camera direction
---------
Co-authored-by: Qirui Sun <sunqr0667@126.com>
2025-05-15 19:00:43 -04:00
comfyanonymous
4a9014e201
Hunyuan Custom initial untested implementation. ( #8101 )
2025-05-13 15:53:47 -04:00
comfyanonymous
640c47e7de
Fix torch warning about deprecated function. ( #8075 )
...
Drop support for torch versions below 2.2 on the audio VAEs.
2025-05-12 14:32:01 -04:00
blepping
42da274717
Use normal ComfyUI attention in ACE-Steps model ( #8023 )
...
* Use normal ComfyUI attention in ACE-Steps model
* Let optimized_attention handle output reshape for ACE
2025-05-09 13:51:02 -04:00
comfyanonymous
fd08e39588
Make torchaudio not a hard requirement. ( #7987 )
...
Some platforms can't install it apparently so if it's not there it should
only break models that actually use it.
2025-05-07 21:37:12 -04:00
comfyanonymous
cc33cd3422
Experimental lyrics strength for ACE. ( #7984 )
2025-05-07 19:22:07 -04:00
Benjamin Berman
b6d3f1fb08
Accept workflows from the command line
2025-05-07 14:53:39 -07:00
comfyanonymous
16417b40d9
Initial ACE-Step model implementation. ( #7972 )
2025-05-07 08:33:34 -04:00
comfyanonymous
80a44b97f5
Change lumina to native RMSNorm. ( #7935 )
2025-05-04 06:39:23 -04:00
comfyanonymous
9187a09483
Change cosmos and hydit models to use the native RMSNorm. ( #7934 )
2025-05-04 06:26:20 -04:00
comfyanonymous
3041e5c354
Switch mochi and wan modes to use pytorch RMSNorm. ( #7925 )
...
* Switch genmo model to native RMSNorm.
* Switch WAN to native RMSNorm.
2025-05-03 19:07:55 -04:00
comfyanonymous
aa9d759df3
Switch ltxv to use the pytorch RMSNorm. ( #7897 )
2025-05-01 06:33:42 -04:00
comfyanonymous
08ff5fa08a
Cleanup chroma PR.
2025-04-30 20:57:30 -04:00
Silver
4ca3d84277
Support for Chroma - Flux1 Schnell distilled with CFG ( #7355 )
...
* Upload files for Chroma Implementation
* Remove trailing whitespace
* trim more trailing whitespace..oops
* remove unused imports
* Add supported_inference_dtypes
* Set min_length to 0 and remove attention_mask=True
* Set min_length to 1
* get_mdulations added from blepping and minor changes
* Add lora conversion if statement in lora.py
* Update supported_models.py
* update model_base.py
* add uptream commits
* set modelType.FLOW, will cause beta scheduler to work properly
* Adjust memory usage factor and remove unnecessary code
* fix mistake
* reduce code duplication
* remove unused imports
* refactor for upstream sync
* sync chroma-support with upstream via syncbranch patch
* Update sd.py
* Add Chroma as option for the OptimalStepsScheduler node
2025-04-30 20:57:00 -04:00
comfyanonymous
dbc726f80c
Better vace memory estimation. ( #7875 )
2025-04-29 20:42:00 -04:00
comfyanonymous
83d04717b6
Support HiDream E1 model. ( #7857 )
2025-04-28 15:01:15 -04:00
comfyanonymous
3ab231f01f
Fix issue with WAN VACE implementation. ( #7724 )
2025-04-21 23:36:12 -04:00
comfyanonymous
5d0d4ee98a
Add strength control for vace. ( #7717 )
2025-04-21 19:36:20 -04:00
doctorpangloss
5823497d55
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-04-21 13:14:36 -07:00
comfyanonymous
ce22f687cc
Support for WAN VACE preview model. ( #7711 )
...
* Support for WAN VACE preview model.
* Remove print.
2025-04-21 14:40:29 -04:00
comfyanonymous
dbcfd092a2
Set default context_img_len to 257
2025-04-17 12:42:34 -04:00
comfyanonymous
c14429940f
Support loading WAN FLF model.
2025-04-17 12:04:48 -04:00
comfyanonymous
0d720e4367
Don't hardcode length of context_img in wan code.
2025-04-17 06:25:39 -04:00
comfyanonymous
1fc00ba4b6
Make hidream work with any latent resolution.
2025-04-16 18:34:14 -04:00
comfyanonymous
f00f340a56
Reuse code from flux model.
2025-04-16 17:43:55 -04:00
comfyanonymous
9ad792f927
Basic support for hidream i1 model.
2025-04-15 17:35:05 -04:00
comfyanonymous
8a438115fb
add RMSNorm to comfy.ops
2025-04-14 18:00:33 -04:00
Raphael Walker
89e4ea0175
Add activations_shape info in UNet models ( #7482 )
...
* Add activations_shape info in UNet models
* activations_shape should be a list
2025-04-04 21:27:54 -04:00
doctorpangloss
ffc1912eff
Fix issues with tests
2025-04-04 08:27:33 -07:00
doctorpangloss
040a324346
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-03-29 15:57:24 -07:00
comfyanonymous
e471c726e5
Fallback to pytorch attention if sage attention fails.
2025-03-22 15:45:56 -04:00
comfyanonymous
32ca0805b7
Fix orientation of hunyuan 3d model.
2025-03-19 19:55:24 -04:00
comfyanonymous
11f1b41bab
Initial Hunyuan3Dv2 implementation.
...
Supports the multiview, mini, turbo models and VAEs.
2025-03-19 16:52:58 -04:00
comfyanonymous
3b19fc76e3
Allow disabling pe in flux code for some other models.
2025-03-18 05:09:25 -04:00
comfyanonymous
e8e990d6b8
Cleanup code.
2025-03-16 06:29:12 -04:00
comfyanonymous
a2448fc527
Remove useless code.
2025-03-14 18:10:37 -04:00
comfyanonymous
6a0daa79b6
Make the SkipLayerGuidanceDIT node work on WAN.
2025-03-14 10:55:19 -04:00
FeepingCreature
9c98c6358b
Tolerate missing @torch.library.custom_op ( #7234 )
...
This can happen on Pytorch versions older than 2.4.
2025-03-14 09:51:26 -04:00
FeepingCreature
7aceb9f91c
Add --use-flash-attention flag. ( #7223 )
...
* Add --use-flash-attention flag.
This is useful on AMD systems, as FA builds are still 10% faster than Pytorch cross-attention.
2025-03-14 03:22:41 -04:00
comfyanonymous
9aac21f894
Fix issues with new hunyuan img2vid model and bumb version to v0.3.26
2025-03-09 05:07:22 -04:00
comfyanonymous
7395b0c0d1
Support new hunyuan video i2v model.
...
Use the new "v2 (replace)" guidance type in HunyuanImageToVideo and set
image_interleave to 4 on the "Text Encode Hunyuan Video" node.
2025-03-08 20:34:47 -05:00
comfyanonymous
0952569493
Fix stable cascade VAE on some lowvram machines.
2025-03-08 20:24:04 -05:00
doctorpangloss
3c82be86d1
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-03-05 14:38:50 -08:00
comfyanonymous
369b079ff6
Fix lowvram issue with ltxv vae.
2025-03-05 05:26:08 -05:00
comfyanonymous
93fedd92fe
Support LTXV 0.9.5.
...
Credits: Lightricks team.
2025-03-05 00:13:49 -05:00
comfyanonymous
f4dac8ab6f
Wan code small cleanup.
2025-02-27 07:22:42 -05:00
comfyanonymous
3ea3bc8546
Fix wan issues when prompt length is long.
2025-02-26 20:34:02 -05:00
comfyanonymous
0270a0b41c
Reduce artifacts on Wan by doing the patch embedding in fp32.
2025-02-26 16:59:26 -05:00
comfyanonymous
4bca7367f3
Don't try to use clip_fea on t2v model.
2025-02-26 08:38:09 -05:00
comfyanonymous
fa62287f1f
More code reuse in wan.
...
Fix bug when changing the compute dtype on wan.
2025-02-26 05:22:29 -05:00
comfyanonymous
4ced06b879
WIP support for Wan I2V model.
2025-02-26 01:49:43 -05:00
comfyanonymous
9a66bb972d
Make wan work with all latent resolutions.
...
Cleanup some code.
2025-02-25 19:56:04 -05:00
comfyanonymous
ea0f939df3
Fix issue with wan and other attention implementations.
2025-02-25 19:13:39 -05:00
comfyanonymous
f37551c1d2
Change wan rope implementation to the flux one.
...
Should be more compatible.
2025-02-25 19:11:14 -05:00
comfyanonymous
63023011b9
WIP support for Wan t2v model.
2025-02-25 17:20:35 -05:00
doctorpangloss
048746f58b
Update to 0.3.15 and improve models
...
- Cosmos now fully tested
- Preliminary support for essential Cosmos prompt "upsampler"
- Lumina tests
- Tweaks to language and image resizing nodes
- Fix for #31 all the samplers are now present again
2025-02-24 21:27:15 -08:00
doctorpangloss
693038738a
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-02-24 09:39:26 -08:00
comfyanonymous
96d891cb94
Speedup on some models by not upcasting bfloat16 to float32 on mac.
2025-02-24 05:41:32 -05:00
comfyanonymous
aff16532d4
Remove some useless code.
2025-02-22 04:45:14 -05:00
Jukka Seppänen
acc152b674
Support loading and using SkyReels-V1-Hunyuan-I2V ( #6862 )
...
* Support SkyReels-V1-Hunyuan-I2V
* VAE scaling
* Fix T2V
oops
* Proper latent scaling
2025-02-18 17:06:54 -05:00
comfyanonymous
1cd6cd6080
Disable pytorch attention in VAE for AMD.
2025-02-14 05:42:14 -05:00
HishamC
b124256817
Fix for running via DirectML ( #6542 )
...
* Fix for running via DirectML
Fix DirectML empty image generation issue with Flux1. add CPU fallback for unsupported path. Verified the model works on AMD GPUs
* fix formating
* update casual mask calculation
2025-02-11 17:11:32 -05:00
comfyanonymous
4027466c80
Make lumina model work with any latent resolution.
2025-02-10 00:24:20 -05:00
comfyanonymous
14880e6dba
Remove some useless code.
2025-02-06 05:00:37 -05:00
comfyanonymous
94f21f9301
Upcasting rope to fp32 seems to make no difference in this model.
2025-02-05 04:32:47 -05:00
comfyanonymous
60653004e5
Use regular numbers for rope in lumina model.
2025-02-05 04:17:25 -05:00
comfyanonymous
a57d635c5f
Fix lumina 2 batches.
2025-02-04 21:48:11 -05:00
comfyanonymous
3e880ac709
Fix on python 3.9
2025-02-04 04:20:56 -05:00
comfyanonymous
e5ea112a90
Support Lumina 2 model.
2025-02-04 04:16:30 -05:00
Dr.Lt.Data
0a0df5f136
better guide message for sageattention ( #6634 )
2025-02-02 09:26:47 -05:00
doctorpangloss
95a12f42e2
Fix pylint errors (they were real, as they usually are)
2025-01-28 17:16:15 -08:00
doctorpangloss
a3452f6e6a
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-01-28 13:45:51 -08:00
comfyanonymous
96e2a45193
Remove useless code.
2025-01-23 05:56:23 -05:00
comfyanonymous
fb2ad645a3
Add FluxDisableGuidance node to disable using the guidance embed.
2025-01-20 14:50:24 -05:00
comfyanonymous
0aa2368e46
Fix some cosmos fp8 issues.
2025-01-16 17:45:37 -05:00
comfyanonymous
cca96a85ae
Fix cosmos VAE failing with videos longer than 121 frames.
2025-01-16 16:30:06 -05:00
doctorpangloss
cf3c96e593
Cosmos support
2025-01-16 12:39:05 -08:00
doctorpangloss
631d9e44c6
Merge branch 'master' of github.com:comfyanonymous/ComfyUI
2025-01-16 09:58:02 -08:00
comfyanonymous
31831e6ef1
Code refactor.
2025-01-16 07:23:54 -05:00
comfyanonymous
23289a6a5c
Clean up some debug lines.
2025-01-16 04:24:39 -05:00
comfyanonymous
6320d05696
Slightly lower hunyuan video memory usage.
2025-01-16 00:23:01 -05:00
comfyanonymous
25683b5b02
Lower cosmos diffusion model memory usage.
2025-01-15 23:46:42 -05:00
comfyanonymous
4758fb64b9
Lower cosmos VAE memory usage by a bit.
2025-01-15 22:57:52 -05:00
comfyanonymous
008761166f
Optimize first attention block in cosmos VAE.
2025-01-15 21:48:46 -05:00
comfyanonymous
1f1c7b7b56
Remove useless code.
2025-01-13 03:52:37 -05:00
comfyanonymous
2ff3104f70
WIP support for Nvidia Cosmos 7B and 14B text to world (video) models.
2025-01-10 09:14:16 -05:00
comfyanonymous
129d8908f7
Add argument to skip the output reshaping in the attention functions.
2025-01-10 06:27:37 -05:00
comfyanonymous
ff838657fa
Cleaner handling of attention mask in ltxv model code.
2025-01-09 07:12:03 -05:00