Skip to main content

Bypass Exponential Time Preprocessing: Fast Neural Network Training via Weight-Data Correlation Preprocessing

Publication ,  Conference
Alman, J; Liang, J; Song, Z; Zhang, R; Zhuo, D
Published in: Advances in Neural Information Processing Systems
January 1, 2023

Over the last decade, deep neural networks have transformed our society, and they are already widely applied in various machine learning applications. State-of-the-art deep neural networks are becoming larger in size every year to deliver increasing model accuracy, and as a result, model training consumes substantial computing resources and will only consume more in the future. Using current training methods, in each iteration, to process a data point x ∈ Rd in a layer, we need to spend Θ(md) time to evaluate all the m neurons in the layer. This means processing the entire layer takes Θ(nmd) time for n data points. Recent work [Song, Yang and Zhang, NeurIPS 2021] reduces this time per iteration to o(nmd) but requires exponential time to preprocess either the data or the neural network weights, making it unlikely to have practical usage. In this work, we present a new preprocessing method that simply stores the weight-data correlation in a tree data structure in order to quickly, and dynamically detect which neurons fire at each iteration. Our method requires only O(nmd) time in preprocessing and still achieves o(nmd) time per iteration. We complement our new algorithm with a lower bound, proving that assuming a popular conjecture from complexity theory, one could not substantially speed up our algorithm for dynamic detection of firing neurons.

Duke Scholars

Published In

Advances in Neural Information Processing Systems

ISSN

1049-5258

Publication Date

January 1, 2023

Volume

36

Start / End Page

48110 / 48137

Related Subject Headings

  • 4611 Machine learning
  • 1702 Cognitive Sciences
  • 1701 Psychology
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Alman, J., Liang, J., Song, Z., Zhang, R., & Zhuo, D. (2023). Bypass Exponential Time Preprocessing: Fast Neural Network Training via Weight-Data Correlation Preprocessing. In Advances in Neural Information Processing Systems (Vol. 36, pp. 48110–48137).
Alman, J., J. Liang, Z. Song, R. Zhang, and D. Zhuo. “Bypass Exponential Time Preprocessing: Fast Neural Network Training via Weight-Data Correlation Preprocessing.” In Advances in Neural Information Processing Systems, 36:48110–37, 2023.
Alman J, Liang J, Song Z, Zhang R, Zhuo D. Bypass Exponential Time Preprocessing: Fast Neural Network Training via Weight-Data Correlation Preprocessing. In: Advances in Neural Information Processing Systems. 2023. p. 48110–37.
Alman, J., et al. “Bypass Exponential Time Preprocessing: Fast Neural Network Training via Weight-Data Correlation Preprocessing.” Advances in Neural Information Processing Systems, vol. 36, 2023, pp. 48110–37.
Alman J, Liang J, Song Z, Zhang R, Zhuo D. Bypass Exponential Time Preprocessing: Fast Neural Network Training via Weight-Data Correlation Preprocessing. Advances in Neural Information Processing Systems. 2023. p. 48110–48137.

Published In

Advances in Neural Information Processing Systems

ISSN

1049-5258

Publication Date

January 1, 2023

Volume

36

Start / End Page

48110 / 48137

Related Subject Headings

  • 4611 Machine learning
  • 1702 Cognitive Sciences
  • 1701 Psychology