Papers Using This Method
Covariant Gradient Descent2025-04-07AdamL: A fast adaptive gradient method incorporating loss function2023-12-23AdaPlus: Integrating Nesterov Momentum and Precise Stepsize Adjustment on AdamW Basis2023-09-05On Suppressing Range of Adaptive Stepsizes of Adam to Improve Generalisation Performance2023-02-02Optimization without Backpropagation2022-09-13Moment Centralization based Gradient Descent Optimizers for Convolutional Neural Networks2022-07-19Efficient Adaptive Ensembling for Image Classification2022-06-15A Control Theoretic Framework for Adaptive Gradient Optimizers in Machine Learning2022-06-04A DNN Optimizer that Improves over AdaBelief by Suppression of the Adaptive Stepsize Range2022-03-24AdaFamily: A family of Adam-like adaptive gradient methods2022-03-03Improvements to short-term weather prediction with recurrent-convolutional networks2021-11-11Effectiveness of Optimization Algorithms in Deep Image Classification2021-10-04Using Self-Supervised Auxiliary Tasks to Improve Fine-Grained Facial Representation2021-05-13FastAdaBelief: Improving Convergence Rate for Belief-based Adaptive Optimizers by Exploiting Strong Convexity2021-04-28Adversarial example generation with AdaBelief Optimizer and Crop Invariance2021-02-07Generalizing Adversarial Examples by AdaBelief Optimizer2021-01-25FCM-RDpA: TSK Fuzzy Regression Model Construction Using Fuzzy C-Means Clustering, Regularization, DropRule, and Powerball AdaBelief2020-11-30EAdam Optimizer: How $ε$ Impact Adam2020-11-04AdaBelief Optimizer: Adapting Stepsizes by the Belief in Observed Gradients2020-10-15