Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

eta parameters are trainable #1

Open
backpropper opened this issue Jun 20, 2018 · 1 comment
Open

eta parameters are trainable #1

backpropper opened this issue Jun 20, 2018 · 1 comment

Comments

@backpropper
Copy link

All etas are trainable parameters. You cannot set requires_grad=False for them.

@pemami4911
Copy link
Owner

Right! When I was playing around with the author's TF implementation, I noticed that the eta's were staying very close to 1. So for both the TF implementation and my PyTorch implementation, I experimented with fixing them at 1. From the plot in the README, on the simple binarized MNIST problem, the author's TF implementation of REBAR is still able to do quite well without optimizing the etas. So, I think the issue with my code is something else. I'll add the objective for optimizing the etas to my code soon, though. Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants