Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Turing architecture error on Nvidia Quadro T1000 #1230

Open
Tortoise17 opened this issue Sep 15, 2024 · 2 comments
Open

Turing architecture error on Nvidia Quadro T1000 #1230

Tortoise17 opened this issue Sep 15, 2024 · 2 comments

Comments

@Tortoise17
Copy link

I am facing this error

RuntimeError: FlashAttention only supports Ampere GPUs or newer.

while architecture is Turing. Is there any tip to resolve it?
GPU is NVIDIA T1000.

kindly help.

@Tortoise17 Tortoise17 changed the title Turing architecture error on Nvidia T1000 Turing architecture error on Nvidia Quadro T1000 Sep 16, 2024
@Carnyzzle
Copy link

Carnyzzle commented Sep 16, 2024

Flash attention 1.x supports Turing, Flash attention 2.x doesn't support Turing as of right now.

@Tortoise17
Copy link
Author

@Carnyzzle Thank you. I downgraded to flash_attn 1.xx and still same error. If you can specifically mention which version can get it resolved would be a great help

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants