You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
| AdaBelief |*Adapting Step-sizes by the Belief in Observed Gradients*|[github](https://github.com/juntang-zhuang/Adabelief-Optimizer)|<https://arxiv.org/abs/2010.07468>|[cite](https://ui.adsabs.harvard.edu/abs/2020arXiv201007468Z/exportcitation)|
@@ -182,6 +192,16 @@ from pytorch_optimizer import get_supported_lr_schedulers
| Explore-Exploit |*Wide-minima Density Hypothesis and the Explore-Exploit Learning Rate Schedule*||<https://arxiv.org/abs/2003.03977>|[cite](https://ui.adsabs.harvard.edu/abs/2020arXiv200303977I/exportcitation)|
@@ -199,6 +219,16 @@ from pytorch_optimizer import get_supported_loss_functions
| Label Smoothing |*Rethinking the Inception Architecture for Computer Vision*||<https://arxiv.org/abs/1512.00567>|[cite](https://ui.adsabs.harvard.edu/abs/2015arXiv151200567S/exportcitation)|
0 commit comments