-
Notifications
You must be signed in to change notification settings - Fork 117
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How should I use multiple GPUs? #15
Comments
The original Meta's repo works with A100s as well. |
It stipulates that 65B must use 8GPU |
I'm also trying to figure out how to run with 2 gpus |
how to run example-chat.py with 8A100? |
If you succeed, please tell me |
Reshard the model for the number of GPUs you have: Resharding the larger models to one file can improve load times in general. |
@Chting @wgimperial @fmeres now you may try to run HF version on a more than one GPU's. |
Thank you very much |
I'm testing 65B. One A100 is too slow. I want to use two or four
The text was updated successfully, but these errors were encountered: