-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Tips and Tricks for training classification convolutional neural networks #11
Comments
The highest prio to me is:
@tomMoral @pierreablin wdyt? |
RE LR scheduling: there exists a big difference in how TF and PL implement it. I might just go with the PL way of doing it, since it's the least flexible. |
RE LR scheduling / Weight Decay: I am not sure what is the canonical way of updating the weight decay given the lr schedule. In TF, it is specified that it should be updated, and in this case manually:
But in PL or torch, I didn't see any mention of this update, so it might not be used. I am going to verify this. EDITOk so the problem with WD is actually the following, and I understood it reading the original decoupled weight decay paper as well as the docs of Adam and AdamW in PyTorch.
I will call both types clearly in the solvers. |
Data augmentation:
Regularization:
Learning rate:
Modeling (to me these ones are out of our scope):
Other:
The text was updated successfully, but these errors were encountered: