-
Notifications
You must be signed in to change notification settings - Fork 10
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Inquiry: Scaling the Lr #2
Comments
Here: Line 94 in 60bd564
Contrary to pytorch optimizer, the SGLD optimizer takes in an optional lr as argument! |
Oh thanks, and how do you scale the learning rate with respect to the number of data points ? Looking at ChunyuanLI implementation (ChunyuanLI/pSGLD#2) and at gmarceaucaron implementation (https://github.com/gmarceaucaron/natural-langevin-dynamics-for-neural-networks) they scale the langevin noise with respect to the number of data points in the training (Ntrain) or the square root of Ntrain to allow convergence. Do you do something similar ? |
@rb876 and I discusssing this by email - closing the issue! |
Hi,
it is not clear to me how the Lr is scaled throughout the learning.
Many Thanks
The text was updated successfully, but these errors were encountered: