You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello, I attempted to replicate the results from the paper using the specified settings. According to the paper, the experiments were conducted on a single server with 8 V100 GPUs, and the total batch size was 32. Consequently, the batch size for each GPU should be 4. However, when I used this value, the training consistently failed.
Could you please provide the training scripts that were used in the paper? I would greatly appreciate your assistance.
The text was updated successfully, but these errors were encountered:
Hello, I attempted to replicate the results from the paper using the specified settings. According to the paper, the experiments were conducted on a single server with 8 V100 GPUs, and the total batch size was 32. Consequently, the batch size for each GPU should be 4. However, when I used this value, the training consistently failed.
Could you please provide the training scripts that were used in the paper? I would greatly appreciate your assistance.
The text was updated successfully, but these errors were encountered: