You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I think the kernels are unified. Why is fp8 enabled for mha_fwd but not for mha_varlen_fwd? What's the blocker now?
I'm willing to help and contribute if it's not coming recently.
Update - I tired to enable fp8 for mha_varlen_fwd and I got CUDA illegal memory access error.
Thanks!
The text was updated successfully, but these errors were encountered:
goldhuang
changed the title
Why is fp8 not enabled for mha_varlen_fwd?
fp8 not enabled for mha_varlen_fwd
Sep 17, 2024
I created an issue earlier. #1157.
https://github.com/Dao-AILab/flash-attention/blob/main/hopper/flash_api.cpp#L447.
I think the kernels are unified. Why is fp8 enabled for mha_fwd but not for mha_varlen_fwd? What's the blocker now?
I'm willing to help and contribute if it's not coming recently.
Update - I tired to enable fp8 for mha_varlen_fwd and I got CUDA illegal memory access error.
Thanks!
The text was updated successfully, but these errors were encountered: