-
Notifications
You must be signed in to change notification settings - Fork 144
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training slows down #25
Comments
Do the batch dimensions change through time?? |
Hi, the batch did not change over time. The time dimension in my case is 51, and the batch size is 60, meaning 60 trajectories and each trajectory's length is 51. |
@papoudakis Hi, I tried to debug the code with MLP+IPPO (in the melting pot, I used CNN) for training the melting pot, I found the time cost per update increased as shown below (3 seeds). For each time step, normally the time cost is 0.9, but there are some updates whose time cost could be larger than 0.9. Such a pattern is frequent. In my previous figure, as you can see, such a pattern is not normal. I just wonder are there any potential issues with the code of the IPPO learner? By the way, I did not change the code of the IPPO learner, what I just changed is the agent's network for the training of the melting pot. The following is the time cost before the 30M steps of the above figure. Do you have any clue? |
Hi, this is an issue of sacred, IDSIA/sacred#877 |
Hi, I am using epymarl to train melting pot, I replaced the rnn+mlp network with rnn+cnn. I found in IPPO and IA2C that the time cost for learning each batch increases over time shown below. I tried my best to debug it and I found it very hard to find the reason. I even used
th.cuda.empty_cache()
andth.cuda.synchronize(device=th.device("cuda"))
but no help. The following figure shows the average time cost of the past 10 updates.Did you find also found such an issue?
The text was updated successfully, but these errors were encountered: