-
Notifications
You must be signed in to change notification settings - Fork 6.2k
[core] support attention backends for LTX #12021
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks! Just some minor asks in the refactoring
hidden_states = hidden_states.to(query.dtype) | ||
|
||
hidden_states = attn.to_out[0](hidden_states) | ||
hidden_states = attn.to_out[1](hidden_states) | ||
return hidden_states | ||
|
||
|
||
class LTXAttention(torch.nn.Module, AttentionModuleMixin): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
super clean!
Co-authored-by: Aryan <aryan@huggingface.co>
Thanks @a-r-r-o-w! Have you started Wan already? If so, cool! If not, can I start and test your parallelism PR a bit with that? |
Wan attention backend support has already been merged in #11918. For testing parallelism, I've tested most of the implementations (even some outside diffusers) to validate the soundness of going forth with CP-plans. Some plans are available here: https://github.com/huggingface/finetrainers/blob/f476c3717da6cbfb1070505a99ee989426b46d9c/finetrainers/models/_metadata/transformer.py#L68 |
Cool! Once this gets matured a bit would love documenting some benchmakrs across the board. |
What does this PR do?
Working code:
Unfold
Output
output.mp4
Regarding the output:
main
.