Skip to main content

HeteRSGD: Tackling Heterogeneous Sampling Costs via Optimal Reweighted Stochastic Gradient Descent

Publication ,  Conference
Chen, Z; Lu, J; Qian, H; Wang, X; Yin, W
Published in: Proceedings of Machine Learning Research
January 1, 2023

One implicit assumption in current stochastic gradient descent (SGD) algorithms is the identical cost for sampling each component function of the finite-sum objective. However, there are applications where the costs differ substantially, for which SGD schemes with uniform sampling invoke a high sampling load. We investigate the use of importance sampling (IS) as a cost saver in this setting, in contrast to its traditional use for variance reduction. The key ingredient is a novel efficiency metric for IS that advocates low sampling costs while penalizing high gradient variances. We then propose HeteRSGD, an SGD scheme that performs gradient sampling according to optimal probability weights stipulated by the metric, and establish theories on its optimal asymptotic and finite-time convergence rates among all possible IS-based SGD schemes. We show that the relative efficiency gain of HeteRSGD can be arbitrarily large regardless of the problem dimension and number of components. Our theoretical results are validated numerically for both convex and nonconvex problems.

Duke Scholars

Published In

Proceedings of Machine Learning Research

EISSN

2640-3498

Publication Date

January 1, 2023

Volume

206

Start / End Page

10732 / 10781
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Chen, Z., Lu, J., Qian, H., Wang, X., & Yin, W. (2023). HeteRSGD: Tackling Heterogeneous Sampling Costs via Optimal Reweighted Stochastic Gradient Descent. In Proceedings of Machine Learning Research (Vol. 206, pp. 10732–10781).
Chen, Z., J. Lu, H. Qian, X. Wang, and W. Yin. “HeteRSGD: Tackling Heterogeneous Sampling Costs via Optimal Reweighted Stochastic Gradient Descent.” In Proceedings of Machine Learning Research, 206:10732–81, 2023.
Chen Z, Lu J, Qian H, Wang X, Yin W. HeteRSGD: Tackling Heterogeneous Sampling Costs via Optimal Reweighted Stochastic Gradient Descent. In: Proceedings of Machine Learning Research. 2023. p. 10732–81.
Chen, Z., et al. “HeteRSGD: Tackling Heterogeneous Sampling Costs via Optimal Reweighted Stochastic Gradient Descent.” Proceedings of Machine Learning Research, vol. 206, 2023, pp. 10732–81.
Chen Z, Lu J, Qian H, Wang X, Yin W. HeteRSGD: Tackling Heterogeneous Sampling Costs via Optimal Reweighted Stochastic Gradient Descent. Proceedings of Machine Learning Research. 2023. p. 10732–10781.

Published In

Proceedings of Machine Learning Research

EISSN

2640-3498

Publication Date

January 1, 2023

Volume

206

Start / End Page

10732 / 10781