We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hi Team, _flash_attn_forward now supports fp8 but _flash_attn_varlen_forward does not support fp8 yet https://github.com/Dao-AILab/flash-attention/blob/main/hopper/flash_api.cpp#L440. I would like to ask if there are any plans to implement support for _flash_attn_varlen_forward using FP8.
_flash_attn_forward
_flash_attn_varlen_forward
Thank you!
The text was updated successfully, but these errors were encountered:
No branches or pull requests
Hi Team,
_flash_attn_forward
now supports fp8 but_flash_attn_varlen_forward
does not support fp8 yet https://github.com/Dao-AILab/flash-attention/blob/main/hopper/flash_api.cpp#L440.I would like to ask if there are any plans to implement support for _flash_attn_varlen_forward using FP8.
Thank you!
The text was updated successfully, but these errors were encountered: