You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, I wanna a faster transformer implemention in pytorch, and I found one in pytorch code ,whose path is pytorch/aten/src/Aten/native/transformers/cuda/, and it needs support USE_FLASH_ATTENTION compile. Furtherly I found some asm ptx code in utils.h, and amd-pytorch doesn't support it until now. So do you have any plan to support this feature?
The text was updated successfully, but these errors were encountered:
xbcReal
changed the title
when would you support USE_FLASH_ATTENTION compile?
When would you support USE_FLASH_ATTENTION compile?
Jul 7, 2023
🚀 The feature, motivation and pitch
Hi, I wanna a faster transformer implemention in pytorch, and I found one in pytorch code ,whose path is pytorch/aten/src/Aten/native/transformers/cuda/, and it needs support USE_FLASH_ATTENTION compile. Furtherly I found some asm ptx code in utils.h, and amd-pytorch doesn't support it until now. So do you have any plan to support this feature?
The text was updated successfully, but these errors were encountered: