-
Notifications
You must be signed in to change notification settings - Fork 2.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feature : Enable Intel GPU/XPU finetuning and inference #116
Conversation
@abhilash1910 thanks the PR, very interesting, would love to learn more about how we can test Intel GPUs/XPUs. I sent you a message on LD. |
@abhilash1910 Thanks for submitting this PR. Please attach the logs for the tests for verification that things are working on the Intel GPUs with the changes. We don't have a mechanism to test this. Can you also please check if torch.set_default_device api can be used to simply the device setting logic for the XPU devices? |
@chauhang sure I will attach the logs; also as discussed with @HamidShojanazeri there is a plan to support device for integration testing for any future PRs from Intel side. I am closing on it this week internally , and eventually would like to discuss on the next steps. |
@chauhang @HamidShojanazeri logs for distributed finetuning on 8 ranks : |
Thanks very much @abhilash1910, I am wondering if there is a way to add support/ get intel GPU memory stats as well the way we are getting for nvidia. |
@HamidShojanazeri yes the memory stats are also added , this can be found in the new log here: |
Upstream from main
What does this PR do?
Motivation: Thanks for creating this repository. At Intel, we are using this repository for finetuning and inference for our GPUs /XPU (Ipex backend) and hence the reason for this PR.
This change would help us run LLama v2 without issues on our GPU devices. (Since accelerate/peft already has support of XPUs enabled)
Fixes # (issue) -NA
Feature/Issue validation/testing
Feature for finetuning and inference tested on our Data Centre GPUs with the change
for CLA purposes this would be on behalf of Intel.
Before submitting
Pull Request section?
to it if that's the case.
Thanks for contributing 🎉!