Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Learning Rate Setting #16

Open
mr-Mojo opened this issue Nov 5, 2019 · 1 comment
Open

Learning Rate Setting #16

mr-Mojo opened this issue Nov 5, 2019 · 1 comment

Comments

@mr-Mojo
Copy link

mr-Mojo commented Nov 5, 2019

Hi, I am experimenting with the NIMA implementation for a scientfic project! In your Readme, you say that "The learning rate setting differs from the original paper. I can't seem to get the model to converge with momentum SGD using an lr of 3e-7 for the conv base and 3e-6 for the dense block.". Which settings did you use? The defaults in argparser are set to the 3e-7, 3e-6, so that's why I was wondering!

Thank you :)

@yunxiaoshi
Copy link
Owner

0.005 for the dense block and 0.0005 for the conv base gives you a reasonable result. But I suspect it is not optimal, maybe do a grid/random search

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants