ComfyUI/comfy/ldm/wan
rattus128 653ceab414
Reduce Peak WAN inference VRAM usage - part II (#10062)
* flux: math: Use _addcmul to avoid expensive VRAM intermediate

The rope process can be the VRAM peak and this intermediate
for the addition result before releasing the original can OOM.
addcmul_ it.

* wan: Delete the self attention before cross attention

This saves VRAM when the cross attention and FFN are in play as the
VRAM peak.
2025-09-27 18:14:16 -04:00
..
model_animate.py Update WanAnimateToVideo to more easily extend videos. (#9959) 2025-09-19 18:48:56 -04:00
model.py Reduce Peak WAN inference VRAM usage - part II (#10062) 2025-09-27 18:14:16 -04:00
vae2_2.py Tiny wan vae optimizations. (#9136) 2025-08-01 05:25:38 -04:00
vae.py Tiny wan vae optimizations. (#9136) 2025-08-01 05:25:38 -04:00