Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

FA 2.4.2 is falling unitest on A6000 and A5880 #1409

Open
BoxiangW opened this issue Dec 23, 2024 · 0 comments
Open

FA 2.4.2 is falling unitest on A6000 and A5880 #1409

BoxiangW opened this issue Dec 23, 2024 · 0 comments

Comments

@BoxiangW
Copy link
Contributor

I am using NVIDIA's 24.10-py3 container. This is how to reproduce:

git clone https://github.com/Dao-AILab/flash-attention.git
Verify the installed version of python -c 'import flash_attn; print(flash_attn.__version__)'
cd flash-attention && git checkout v2.4.2 && cd .. # for example, make sure you are not in the flash-attention dir when you run the test
py.test flash-attention/tests/test_flash_attn.py::test_flash_attn_qkvpacked[0.17-2048-160-True-True-False-False-dtype0]

I tested that this only happens on A5880 and A6000. Could be affecting A100 as well. H100 is not affected.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant