-
Notifications
You must be signed in to change notification settings - Fork 6k
"joint_attention_kwargs" don't pass the parameters to AttentionProcessor #8855
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
cc: @sayakpaul could you take a look here. |
Hi. We can make that happen perhaps with an IPAdapterSD3Processor but for now
and here
Does this make sense? |
Is there any chance this can be changed? Not being able to pass parameters (like an attention mask) to the attention processor makes things really difficult. If this can't or won't be changed, I suggest at least changing the doc string. It currently says that the kwargs are passed to the processor:
|
We usually update that when there is a use case. If you can showcase a viable use case, we will definitely welcome the change. For SD3 (where
You're right. Thanks for pointing that out. Could you open a PR for that? Since you know the solution, we want to honor your contribution via a PR :) |
I want to train my ipadapter in sd3. I noticed that joint_attention_kwargs is used in passing the parameters to AttentionProcessor (https://github.com/huggingface/diffusers/blob/main/src/diffusers/models/transformers/transformer_sd3.py#L309). But it seems joint_attention_kwargs is not passed to transformer_blocks. Is my usage method wrong or this parameter has not been developed yet? Thank you!
The text was updated successfully, but these errors were encountered: