1 code implementation • 7 Sep 2021 • Anirudh Maiya, Inumella Sricharan, Anshuman Pandey, Srinivas K. S
To eliminate the dependency of learning rate schedulers, adaptive gradient optimizers such as AdaGrad, AdaDelta, RMSProp, Adam employ a parameter-wise scaling term for learning rate which is a function of the gradient itself.
1 code implementation • 9 Feb 2021 • Anuj Tambwekar, Anirudh Maiya, Soma Dhavala, Snehanshu Saha
We quantify the uncertainty of the class probabilities in terms of prediction intervals, and develop individualized confidence scores that can be used to decide whether a prediction is reliable or not at scoring time.