Skip to content

Support TEI on AMD GPUs #108

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
japarada opened this issue Dec 13, 2023 · 9 comments
Open

Support TEI on AMD GPUs #108

japarada opened this issue Dec 13, 2023 · 9 comments

Comments

@japarada
Copy link

japarada commented Dec 13, 2023

Feature request

Are there active plans to add support for generating embedding using AMD GPUs?

Motivation

AMD and Hugging Face are currently engaged in an ongoing collaboration to ensure compatibility of Hugging Face transformers with AMD ROCm and hardware platforms. Providing support for HF TEI toolkit would add a compelling alternative for deploying and serving open-source text embeddings and sequence classification models.

Your contribution

Work with other on PRs to integrate changes to support AMD GPUs.

@OlivierDehaene
Copy link
Contributor

Indeed it would be interesting to support more backends.
Adding support for AMD GPUs will happen in Candle tough, not in TEI directly.

@japarada
Copy link
Author

@OlivierDehaene Thanks for the response. Does TEI support python backend instead of Candle? I see https://github.com/huggingface/text-embeddings-inference/tree/d05c949c1234786c15b675f4419776a417519583/backends/python? Is this python code only for the grpc server implementation?

@OlivierDehaene
Copy link
Contributor

This backend is an example of how you would go about adding other backends to TEI. It is currently out-dated but could be updated to support AMD.

@japarada
Copy link
Author

Are there any active effort support to make python backend fully functional?
What models can be run at this moment?
Thank you for all the answers.

@hvico
Copy link

hvico commented May 20, 2024

Hi! Is there any WiP regarding ROCm support for TEI? Thanks!

@dcbark01
Copy link

Hello? Is there any WiP regarding ROCm support for TEI? Thanks!

I'm interested in this as well. May start working on it myself, but don't want to duplicate efforts if it's already in the works.

@fxmarty
Copy link
Contributor

fxmarty commented Jun 19, 2024

Hi, there is some progress in #293. Would you mind sharing which AMD GPUs you are using? Thank you!

@dcbark01
Copy link

Currently using MI250s.

@puppetm4st3r
Copy link

Hi, any news on this, is there a way to serve the models with amd gpus on TEI?

regards!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

6 participants