mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2026-01-27 06:40:16 +08:00
Enable sage attention autodetection
This commit is contained in:
parent
388dad67d5
commit
99f0fa8b50
@ -420,7 +420,7 @@ def attention_pytorch(q, k, v, heads, mask=None, attn_precision=None, skip_resha
|
|||||||
|
|
||||||
|
|
||||||
@pytorch_style_decl
|
@pytorch_style_decl
|
||||||
def attention_sagemaker(q, k, v, heads, mask=None, attn_precision=None, skip_reshape=False):
|
def attention_sageattn(q, k, v, heads, mask=None, attn_precision=None, skip_reshape=False):
|
||||||
return sageattn(q, k, v, attn_mask=mask, dropout_p=0.0, is_causal=False) # pylint: disable=possibly-used-before-assignment
|
return sageattn(q, k, v, attn_mask=mask, dropout_p=0.0, is_causal=False) # pylint: disable=possibly-used-before-assignment
|
||||||
|
|
||||||
|
|
||||||
@ -431,7 +431,10 @@ def attention_flash_attn(q, k, v, heads, mask=None, attn_precision=None, skip_re
|
|||||||
|
|
||||||
optimized_attention = attention_basic
|
optimized_attention = attention_basic
|
||||||
|
|
||||||
if model_management.xformers_enabled():
|
if model_management.sage_attention_enabled():
|
||||||
|
logging.debug("Using sage attention")
|
||||||
|
optimized_attention = attention_sageattn
|
||||||
|
elif model_management.xformers_enabled():
|
||||||
logging.debug("Using xformers cross attention")
|
logging.debug("Using xformers cross attention")
|
||||||
optimized_attention = attention_xformers
|
optimized_attention = attention_xformers
|
||||||
elif model_management.pytorch_attention_enabled():
|
elif model_management.pytorch_attention_enabled():
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user