Skip to main content

Beyond lazy training for over-parameterized tensor decomposition

Publication ,  Journal Article
Wang, X; Wu, C; Lee, JD; Ma, T; Ge, R
Published in: Advances in Neural Information Processing Systems
January 1, 2020

Over-parametrization is an important technique in training neural networks. In both theory and practice, training a larger network allows the optimization algorithm to avoid bad local optimal solutions. In this paper we study a closely related tensor decomposition problem: given an l-th order tensor in (Rd)?l of rank r (where r « d), can variants of gradient descent find a rank m decomposition where m > r? We show that in a lazy training regime (similar to the NTK regime for neural networks) one needs at least m = ?(dl-1), while a variant of gradient descent can find an approximate tensor when m = O*(r2.5l log d). Our results show that gradient descent on over-parametrized objective could go beyond the lazy training regime and utilize certain low-rank structure in the data.

Duke Scholars

Published In

Advances in Neural Information Processing Systems

ISSN

1049-5258

Publication Date

January 1, 2020

Volume

2020-December

Related Subject Headings

  • 4611 Machine learning
  • 1702 Cognitive Sciences
  • 1701 Psychology
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Wang, X., Wu, C., Lee, J. D., Ma, T., & Ge, R. (2020). Beyond lazy training for over-parameterized tensor decomposition. Advances in Neural Information Processing Systems, 2020-December.
Wang, X., C. Wu, J. D. Lee, T. Ma, and R. Ge. “Beyond lazy training for over-parameterized tensor decomposition.” Advances in Neural Information Processing Systems 2020-December (January 1, 2020).
Wang X, Wu C, Lee JD, Ma T, Ge R. Beyond lazy training for over-parameterized tensor decomposition. Advances in Neural Information Processing Systems. 2020 Jan 1;2020-December.
Wang, X., et al. “Beyond lazy training for over-parameterized tensor decomposition.” Advances in Neural Information Processing Systems, vol. 2020-December, Jan. 2020.
Wang X, Wu C, Lee JD, Ma T, Ge R. Beyond lazy training for over-parameterized tensor decomposition. Advances in Neural Information Processing Systems. 2020 Jan 1;2020-December.

Published In

Advances in Neural Information Processing Systems

ISSN

1049-5258

Publication Date

January 1, 2020

Volume

2020-December

Related Subject Headings

  • 4611 Machine learning
  • 1702 Cognitive Sciences
  • 1701 Psychology