-
Notifications
You must be signed in to change notification settings - Fork 2.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[QUESTION] How to enable ZeRO 2/3 stages ? #1156
Comments
I responded to this on #589. |
please convert this issue to feature request for ZeRO 2/3 |
i think article,https://www.deepspeed.ai/tutorials/megatron/, is useful. |
@carolove Thanks for the inputs, i am familiar with deepspeed framework to enable all ZeRO stages. here query is regarding enabling ZeRO in this repo natively. |
I also look for such example~. |
megatron-lm now has its own zero-1 (it is called distributed optimizer in this project), but if u are more familiar with deepspeed, then how bout using deepspeed-megatron, @polisettyvarma ? |
Thank you @SeunghyunSEO for your inputs. Yes Megatron-DeepSpeed repo can be used but it's not up to date with Megatron-LM. I agree on Zero > 1 is not compatible with PP. |
We should have PyTorch FSDP support compatible with TP in the next couple of weeks. |
Thank you @deepakn94 for sharing this information. |
How to enable ZeRO 2/3 stages ?
similar to #589
The text was updated successfully, but these errors were encountered: