Skip to content

Conversation

@dxqb
Copy link

@dxqb dxqb commented Dec 26, 2025

What does this PR do?

Using an attention backend (https://huggingface.co/docs/diffusers/main/optimization/attention_backends) with a model that passes attention masks yields incorrect results.

This is already checked in parallel backends...

raise ValueError("`attn_mask` is not yet supported for flash-attn 2.")

...but not yet in the regular ones.

This PR changes that.
Fixes #12605

Who can review?

@yiyixuxu and @asomoza
CC @zzlol63 @tolgacangoz

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

1 participant