Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature]: FlashAttention 3 support #6348

Closed
orellavie1212 opened this issue Jul 11, 2024 · 17 comments
Closed

[Feature]: FlashAttention 3 support #6348

orellavie1212 opened this issue Jul 11, 2024 · 17 comments
Labels
feature request help wanted Extra attention is needed

Comments

@orellavie1212
Copy link
Contributor

🚀 The feature, motivation and pitch

As you know, FA3 promises 1.5x~ improvements
Dao-AILab/flash-attention@7ef2484

Alternatives

No response

Additional context

No response

@simon-mo
Copy link
Collaborator

simon-mo commented Jul 11, 2024

Yes. Actively looking.

Update: It seem Dao-AILab/flash-attention#1268 is merged. The integration is now ready for testing.

@Navanit-git
Copy link

any updates??

@Ushnish-Sengupta
Copy link

This would be amazing. FA3 is apparently part of the reason behind Together Inference Engine leapfrogging vllm by a lot!

@ehartford
Copy link

Very interested in this

@freddifederica
Copy link

+1

2 similar comments
@CambioML
Copy link

+1

@badrisnps
Copy link

+1

@simon-mo
Copy link
Collaborator

A small progress update here. @felixzhu555 did some testing. FA3 is mostly a training kernel at the moment and only benefit extremely large batch size with long context.

@tmoon
Copy link

tmoon commented Nov 5, 2024

Any further update on this?

@yuvalkk2002
Copy link

+1

@simon-mo simon-mo added the help wanted Extra attention is needed label Nov 21, 2024
@hitcoogle
Copy link

+1

5 similar comments
@taegeonum
Copy link

+1

@Isaac4real
Copy link

+1

@HelenaSak
Copy link

+1

@celalettinbilgen2
Copy link

+1

@leo-pony
Copy link

+1

@hmellor
Copy link
Collaborator

hmellor commented Feb 21, 2025

#12093

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature request help wanted Extra attention is needed
Projects
None yet
Development

No branches or pull requests