ComfyUI/comfy_execution
Rattus f3f526fcd3 Implement RAM Pressure cache
Implement a cache sensitive to RAM pressure. When RAM headroom drops
down below a certain threshold, evict RAM-expensive nodes from the
cache.

Models and tensors are measured directly for RAM usage. An OOM score
is then computed based on the RAM usage of the node.

Note the due to indirection through shared objects (like a model
patcher), multiple nodes can account the same RAM as their individual
usage. The intent is this will free chains of nodes particularly
model loaders and associate loras as they all score similar and are
sorted in close to each other.

Has a bias towards unloading model nodes mid flow while being able
to keep results like text encodings and VAE.
2025-10-24 00:51:08 +10:00
..
caching.py Implement RAM Pressure cache 2025-10-24 00:51:08 +10:00
graph_utils.py V3 Node Schema Definition - initial (#8656) 2025-07-31 18:02:12 -04:00
graph.py execution: Roll the UI cache into the outputs 2025-10-23 21:28:38 +10:00
progress.py Fix progress update crossover between users (#9706) 2025-09-04 19:13:28 -04:00
utils.py Support for async node functions (#8830) 2025-07-10 14:46:19 -04:00
validation.py Reland union type (#5900) 2024-12-04 15:12:10 -05:00