mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2026-02-15 16:02:32 +08:00
xpu-workflows-opt
This commit is contained in:
parent
1c10b33f9b
commit
3dab05de8e
@ -493,7 +493,16 @@ def attention_pytorch(q, k, v, heads, mask=None, attn_precision=None, skip_resha
|
||||
mask = mask.unsqueeze(1)
|
||||
|
||||
if SDP_BATCH_LIMIT >= b:
|
||||
out = comfy.ops.scaled_dot_product_attention(q, k, v, attn_mask=mask, dropout_p=0.0, is_causal=False)
|
||||
# out = comfy.ops.scaled_dot_product_attention(q, k, v, attn_mask=mask, dropout_p=0.0, is_causal=False)
|
||||
if mask is None:
|
||||
k = k.contiguous()
|
||||
v = v.contiguous()
|
||||
import sdpa_kernels
|
||||
out = sdpa_kernels.sdp_xmx(q, k, v, 1 / math.sqrt(dim_head), 0)
|
||||
# print("here:", out.isnan().count_nonzero(), out.isinf().count_nonzero())
|
||||
# breakpoint()
|
||||
else:
|
||||
out = comfy.ops.scaled_dot_product_attention(q, k, v, attn_mask=mask, dropout_p=0.0, is_causal=False)
|
||||
if not skip_output_reshape:
|
||||
out = (
|
||||
out.transpose(1, 2).reshape(b, -1, heads * dim_head)
|
||||
|
||||
1
user/default/workflows/image_qwen_image_edit_2509.json
Normal file
1
user/default/workflows/image_qwen_image_edit_2509.json
Normal file
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
Loading…
Reference in New Issue
Block a user