Skip to main content

FastAdaBelief: Improving Convergence Rate for Belief-Based Adaptive Optimizers by Exploiting Strong Convexity.

Publication ,  Journal Article
Zhou, Y; Huang, K; Cheng, C; Wang, X; Hussain, A; Liu, X
Published in: IEEE transactions on neural networks and learning systems
September 2023

AdaBelief, one of the current best optimizers, demonstrates superior generalization ability over the popular Adam algorithm by viewing the exponential moving average of observed gradients. AdaBelief is theoretically appealing in which it has a data-dependent O(√T) regret bound when objective functions are convex, where T is a time horizon. It remains, however, an open problem whether the convergence rate can be further improved without sacrificing its generalization ability. To this end, we make the first attempt in this work and design a novel optimization algorithm called FastAdaBelief that aims to exploit its strong convexity in order to achieve an even faster convergence rate. In particular, by adjusting the step size that better considers strong convexity and prevents fluctuation, our proposed FastAdaBelief demonstrates excellent generalization ability and superior convergence. As an important theoretical contribution, we prove that FastAdaBelief attains a data-dependent O(logT) regret bound, which is substantially lower than AdaBelief in strongly convex cases. On the empirical side, we validate our theoretical analysis with extensive experiments in scenarios of strong convexity and nonconvexity using three popular baseline models. Experimental results are very encouraging: FastAdaBelief converges the quickest in comparison to all mainstream algorithms while maintaining an excellent generalization ability, in cases of both strong convexity or nonconvexity. FastAdaBelief is, thus, posited as a new benchmark model for the research community.

Duke Scholars

Altmetric Attention Stats
Dimensions Citation Stats

Published In

IEEE transactions on neural networks and learning systems

DOI

EISSN

2162-2388

ISSN

2162-237X

Publication Date

September 2023

Volume

34

Issue

9

Start / End Page

6515 / 6529
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Zhou, Y., Huang, K., Cheng, C., Wang, X., Hussain, A., & Liu, X. (2023). FastAdaBelief: Improving Convergence Rate for Belief-Based Adaptive Optimizers by Exploiting Strong Convexity. IEEE Transactions on Neural Networks and Learning Systems, 34(9), 6515–6529. https://doi.org/10.1109/tnnls.2022.3143554
Zhou, Yangfan, Kaizhu Huang, Cheng Cheng, Xuguang Wang, Amir Hussain, and Xin Liu. “FastAdaBelief: Improving Convergence Rate for Belief-Based Adaptive Optimizers by Exploiting Strong Convexity.IEEE Transactions on Neural Networks and Learning Systems 34, no. 9 (September 2023): 6515–29. https://doi.org/10.1109/tnnls.2022.3143554.
Zhou Y, Huang K, Cheng C, Wang X, Hussain A, Liu X. FastAdaBelief: Improving Convergence Rate for Belief-Based Adaptive Optimizers by Exploiting Strong Convexity. IEEE transactions on neural networks and learning systems. 2023 Sep;34(9):6515–29.
Zhou, Yangfan, et al. “FastAdaBelief: Improving Convergence Rate for Belief-Based Adaptive Optimizers by Exploiting Strong Convexity.IEEE Transactions on Neural Networks and Learning Systems, vol. 34, no. 9, Sept. 2023, pp. 6515–29. Epmc, doi:10.1109/tnnls.2022.3143554.
Zhou Y, Huang K, Cheng C, Wang X, Hussain A, Liu X. FastAdaBelief: Improving Convergence Rate for Belief-Based Adaptive Optimizers by Exploiting Strong Convexity. IEEE transactions on neural networks and learning systems. 2023 Sep;34(9):6515–6529.

Published In

IEEE transactions on neural networks and learning systems

DOI

EISSN

2162-2388

ISSN

2162-237X

Publication Date

September 2023

Volume

34

Issue

9

Start / End Page

6515 / 6529