Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about batch size #23

Closed
gyxxyg opened this issue Apr 15, 2024 · 1 comment
Closed

Question about batch size #23

gyxxyg opened this issue Apr 15, 2024 · 1 comment

Comments

@gyxxyg
Copy link

gyxxyg commented Apr 15, 2024

Hello, I attempted to replicate the results from the paper using the specified settings. According to the paper, the experiments were conducted on a single server with 8 V100 GPUs, and the total batch size was 32. Consequently, the batch size for each GPU should be 4. However, when I used this value, the training consistently failed.

Could you please provide the training scripts that were used in the paper? I would greatly appreciate your assistance.

@RenShuhuai-Andy
Copy link
Owner

Hi, what does "The Training Consistently Failed" mean?

Do you mean gpu out-of-memory? If so, please refer to #10 (comment)

@gyxxyg gyxxyg closed this as completed Apr 16, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants