-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add generating batch size #592
base: main
Are you sure you want to change the base?
Conversation
The origin implementation was generating samples one by one, which results in low gpu utility for small model. This commit adds a generating-batch-size config to enable batch generation.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think a better solution is to drop the word train
from the name of the existing argument and use the existing argument for generation. Having two seems needlessly confusing.
@StellaAthena I found that the gpt-neox/eval_tasks/eval_adapter.py Lines 70 to 72 in 1e346ef
Maybe using the |
My suggestion is to rename |
@StellaAthena |
No, I mean just the second rename. The training batch size and the eval batch size for a fixed microbatch size per gpu are typically different due to gradient accumulation. |
@StellaAthena Updated. |
@StellaAthena Hi, could you help me to fix the error raised in CI? Thank you! |
|
The origin implementation was generating samples one by one, which results in low gpu utility for small model. This commit adds a generating-batch-size config to enable batch generation.
Thank you for your time on reviewing this PR :)