Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Inquiry: Scaling the Lr #2

Closed
rb876 opened this issue Jun 20, 2019 · 3 comments
Closed

Inquiry: Scaling the Lr #2

rb876 opened this issue Jun 20, 2019 · 3 comments

Comments

@rb876
Copy link

rb876 commented Jun 20, 2019

Hi,
it is not clear to me how the Lr is scaled throughout the learning.

Many Thanks

@henripal
Copy link
Owner

henripal commented Jun 20, 2019

Here:

optimizer.step(model_desc['lr']/(epoch//model_desc['lr_epoch'] + 1))

Contrary to pytorch optimizer, the SGLD optimizer takes in an optional lr as argument!

@rb876
Copy link
Author

rb876 commented Jun 20, 2019

Oh thanks, and how do you scale the learning rate with respect to the number of data points ? Looking at ChunyuanLI implementation (ChunyuanLI/pSGLD#2) and at gmarceaucaron implementation (https://github.com/gmarceaucaron/natural-langevin-dynamics-for-neural-networks) they scale the langevin noise with respect to the number of data points in the training (Ntrain) or the square root of Ntrain to allow convergence. Do you do something similar ?

@henripal
Copy link
Owner

@rb876 and I discusssing this by email - closing the issue!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants