Skip to main content

DyNNamic: Dynamically Reshaping, High Data-Reuse Accelerator for Compact DNNs

Publication ,  Journal Article
Hanson, E; Li, S; Qian, X; Li, HH; Chen, Y
Published in: IEEE Transactions on Computers
March 1, 2023

Convolutional layers dominate the computation and energy costs of Deep Neural Network (DNN) inference. Recent algorithmic works attempt to reduce these bottlenecks via compact DNN structures and model compression. Likewise, state-of-the-art accelerator designs leverage spatiotemporal characteristics of convolutional layers to reduce data movement overhead and improve throughput. Although both are independently effective at reducing latency and energy costs, combining these approaches does not guarantee cumulative improvements due to inefficient mapping. This inefficiency can be attributed to (1) inflexibility of underlying hardware and (2) inherent reduction of data-reuse opportunities of compact DNN structures. To address these issues, we propose a dynamically reshaping, high data-reuse PE array accelerator, namely DyNNamic. DyNNamic leverages kernel-wise filter decomposition to partition the convolution operation into two compact stages: Shared Kernels Convolution (SKC) and Weighted Accumulation (WA). Because both stages have vastly different dimensions, DyNNamic reshapes its PE array to effectively map the algorithm to the architecture. The architecture then exploits data-reuse opportunities created by the SKC stage, further reducing data movement with negligible overhead. We evaluate our approach on various representative networks and compare against state-of-the-art accelerators. On average, DyNNamic outperforms DianNao by 8.4×8. 4 × and 12.3×12. 3 × in terms of inference energy and latency, respectively.

Duke Scholars

Published In

IEEE Transactions on Computers

DOI

EISSN

1557-9956

ISSN

0018-9340

Publication Date

March 1, 2023

Volume

72

Issue

3

Start / End Page

880 / 892

Related Subject Headings

  • Computer Hardware & Architecture
  • 4606 Distributed computing and systems software
  • 4009 Electronics, sensors and digital hardware
  • 1006 Computer Hardware
  • 0805 Distributed Computing
  • 0803 Computer Software
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Hanson, E., Li, S., Qian, X., Li, H. H., & Chen, Y. (2023). DyNNamic: Dynamically Reshaping, High Data-Reuse Accelerator for Compact DNNs. IEEE Transactions on Computers, 72(3), 880–892. https://doi.org/10.1109/TC.2022.3184272
Hanson, E., S. Li, X. Qian, H. H. Li, and Y. Chen. “DyNNamic: Dynamically Reshaping, High Data-Reuse Accelerator for Compact DNNs.” IEEE Transactions on Computers 72, no. 3 (March 1, 2023): 880–92. https://doi.org/10.1109/TC.2022.3184272.
Hanson E, Li S, Qian X, Li HH, Chen Y. DyNNamic: Dynamically Reshaping, High Data-Reuse Accelerator for Compact DNNs. IEEE Transactions on Computers. 2023 Mar 1;72(3):880–92.
Hanson, E., et al. “DyNNamic: Dynamically Reshaping, High Data-Reuse Accelerator for Compact DNNs.” IEEE Transactions on Computers, vol. 72, no. 3, Mar. 2023, pp. 880–92. Scopus, doi:10.1109/TC.2022.3184272.
Hanson E, Li S, Qian X, Li HH, Chen Y. DyNNamic: Dynamically Reshaping, High Data-Reuse Accelerator for Compact DNNs. IEEE Transactions on Computers. 2023 Mar 1;72(3):880–892.

Published In

IEEE Transactions on Computers

DOI

EISSN

1557-9956

ISSN

0018-9340

Publication Date

March 1, 2023

Volume

72

Issue

3

Start / End Page

880 / 892

Related Subject Headings

  • Computer Hardware & Architecture
  • 4606 Distributed computing and systems software
  • 4009 Electronics, sensors and digital hardware
  • 1006 Computer Hardware
  • 0805 Distributed Computing
  • 0803 Computer Software