-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Specifying batch size on llama2-70B cm automation #190
Comments
@rajesh-s most of the inference submissions are done using Nvidia implementation. In CM we have tried to match the typical batch sizes as in the Nvidia submissions but we haven't tested all of the systems. In the CM run command you can specify For reference implementation I'm not sure if different batch sizes work as many things are hardwired and no one has done any submission using it. |
It would help if the batch sizes are listed atleast on the submissions which I could not find on the results. The CM run command seems to default to the batch size of |
Hi @rajesh-s, sorry for replying late. Have noted the required addition. @arjunsuresh , would it be apt to include in collapsible section or should we give that as a tip since there is a chance of users ignoring the collapsible option. |
I could not find information either in the documentation or in the cm scripts on the batch size that is being used to report the results in the MLCommons database.
1
. Is the cm automation specifying a value different from this?submissions
, should I see nearly the same performance?The text was updated successfully, but these errors were encountered: