Skip to main content

Efficient process-in-memory architecture design for unsupervised GAN-based deep learning using ReRAM

Publication ,  Conference
Chen, F; Song, L; Li, HH
Published in: Proceedings of the ACM Great Lakes Symposium on VLSI, GLSVLSI
May 13, 2019

The ending of Moore's Law makes domain-specific architecture as the future of computing. The most representative is the emergence of various deep learning accelerators. Among the proposed solutions, resistive random access memory (ReRAM) based process-in-memory (PIM) architecture is anticipated as a promising candidate because ReRAM has the capability of both data storage and in-situ computation. However, we found that existing solutions are unable to efficiently support the computational needs required by the training of unsupervised generative adversarial networks (GANs), due to the lack of the following two features: 1) Computation efficiency: GAN utilizes a new operator, called transposed convolution. It inserts massive zeros in its input before a convolution operation, resulting in significant resource under-utilization; 2) Data traffic: The data intensive training process of GANs often incurs structural heavy data traffic as well as frequent massive data swaps. Our research follows the PIM strategy by leveraging the energy-efficiency of ReRAM arrays for vector-matrix multiplication to enhance the performance and energy efficiency. Specifically, we propose a novel computation deformation technique that can skip zero-insertions in transposed convolution for computation efficiency improvement. Moreover, we explore an efficient pipelined training procedure to reduce on-chip memory access. The implementation of related circuits and architecture is also discussed. At the end, we present our perspective on the future trend and opportunities of deep learning accelerators.

Duke Scholars

Published In

Proceedings of the ACM Great Lakes Symposium on VLSI, GLSVLSI

DOI

Publication Date

May 13, 2019

Start / End Page

423 / 428
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Chen, F., Song, L., & Li, H. H. (2019). Efficient process-in-memory architecture design for unsupervised GAN-based deep learning using ReRAM. In Proceedings of the ACM Great Lakes Symposium on VLSI, GLSVLSI (pp. 423–428). https://doi.org/10.1145/3299874.3319482
Chen, F., L. Song, and H. H. Li. “Efficient process-in-memory architecture design for unsupervised GAN-based deep learning using ReRAM.” In Proceedings of the ACM Great Lakes Symposium on VLSI, GLSVLSI, 423–28, 2019. https://doi.org/10.1145/3299874.3319482.
Chen F, Song L, Li HH. Efficient process-in-memory architecture design for unsupervised GAN-based deep learning using ReRAM. In: Proceedings of the ACM Great Lakes Symposium on VLSI, GLSVLSI. 2019. p. 423–8.
Chen, F., et al. “Efficient process-in-memory architecture design for unsupervised GAN-based deep learning using ReRAM.” Proceedings of the ACM Great Lakes Symposium on VLSI, GLSVLSI, 2019, pp. 423–28. Scopus, doi:10.1145/3299874.3319482.
Chen F, Song L, Li HH. Efficient process-in-memory architecture design for unsupervised GAN-based deep learning using ReRAM. Proceedings of the ACM Great Lakes Symposium on VLSI, GLSVLSI. 2019. p. 423–428.

Published In

Proceedings of the ACM Great Lakes Symposium on VLSI, GLSVLSI

DOI

Publication Date

May 13, 2019

Start / End Page

423 / 428