Skip to content

Conversation

@souryadey
Copy link

Adam with weight decay decoupled from gradients, as discussed in 'Fixing Weight Decay Regularization in Adam': https://openreview.net/forum?id=rk6qdGgCZ

This has been shown to give better results in several cases:
https://www.fast.ai/2018/07/02/adam-weight-decay/

Adam with weight decay decoupled from gradients
@pytorchbot pytorchbot added the module: optimizer Related to torch.optim label Jun 24, 2019
@soumith soumith requested a review from vincentqb June 24, 2019 23:04
@soumith
Copy link
Contributor

soumith commented Jun 24, 2019

@vincentqb another AdamW PR for you to review

@soumith soumith added the triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module label Jun 25, 2019
@vincentqb
Copy link
Contributor

Thanks for the implementation. As mentioned in #21250, AdamW landed in master.

@vincentqb vincentqb closed this Jul 2, 2019
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

module: optimizer Related to torch.optim open source triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants