You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Right! When I was playing around with the author's TF implementation, I noticed that the eta's were staying very close to 1. So for both the TF implementation and my PyTorch implementation, I experimented with fixing them at 1. From the plot in the README, on the simple binarized MNIST problem, the author's TF implementation of REBAR is still able to do quite well without optimizing the etas. So, I think the issue with my code is something else. I'll add the objective for optimizing the etas to my code soon, though. Thanks!
All etas are trainable parameters. You cannot set requires_grad=False for them.
The text was updated successfully, but these errors were encountered: