Skip to main content

AutoShrink: A topology-aware NAS for discovering efficient neural architecture

Publication ,  Conference
Zhang, T; Cheng, HP; Li, Z; Yan, F; Huang, C; Li, H; Chen, Y
Published in: AAAI 2020 - 34th AAAI Conference on Artificial Intelligence
January 1, 2020

Resource is an important constraint when deploying Deep Neural Networks (DNNs) on mobile and edge devices. Existing works commonly adopt the cell-based search approach, which limits the flexibility of network patterns in learned cell structures. Moreover, due to the topology-agnostic nature of existing works, including both cell-based and node-based approaches, the search process is time consuming and the performance of found architecture may be sub-optimal. To address these problems, we propose AutoShrink, a topology-aware Neural Architecture Search (NAS) for searching efficient building blocks of neural architectures. Our method is node-based and thus can learn flexible network patterns in cell structures within a topological search space. Directed Acyclic Graphs (DAGs) are used to abstract DNN architectures and progressively optimize the cell structure through edge shrinking. As the search space intrinsically reduces as the edges are progressively shrunk, AutoShrink explores more flexible search space with even less search time. We evaluate AutoShrink on image classification and language tasks by crafting ShrinkCNN and ShrinkRNN models. ShrinkCNN is able to achieve up to 48% parameter reduction and save 34% Multiply-Accumulates (MACs) on ImageNet-1K with comparable accuracy of state-of-the-art (SOTA) models. Specifically, both ShrinkCNN and ShrinkRNN are crafted within 1.5 GPU hours, which is 7.2× and 6.7× faster than the crafting time of SOTA CNN and RNN models, respectively.

Duke Scholars

Published In

AAAI 2020 - 34th AAAI Conference on Artificial Intelligence

Publication Date

January 1, 2020

Start / End Page

6829 / 6836
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Zhang, T., Cheng, H. P., Li, Z., Yan, F., Huang, C., Li, H., & Chen, Y. (2020). AutoShrink: A topology-aware NAS for discovering efficient neural architecture. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 6829–6836).
Zhang, T., H. P. Cheng, Z. Li, F. Yan, C. Huang, H. Li, and Y. Chen. “AutoShrink: A topology-aware NAS for discovering efficient neural architecture.” In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence, 6829–36, 2020.
Zhang T, Cheng HP, Li Z, Yan F, Huang C, Li H, et al. AutoShrink: A topology-aware NAS for discovering efficient neural architecture. In: AAAI 2020 - 34th AAAI Conference on Artificial Intelligence. 2020. p. 6829–36.
Zhang, T., et al. “AutoShrink: A topology-aware NAS for discovering efficient neural architecture.” AAAI 2020 - 34th AAAI Conference on Artificial Intelligence, 2020, pp. 6829–36.
Zhang T, Cheng HP, Li Z, Yan F, Huang C, Li H, Chen Y. AutoShrink: A topology-aware NAS for discovering efficient neural architecture. AAAI 2020 - 34th AAAI Conference on Artificial Intelligence. 2020. p. 6829–6836.

Published In

AAAI 2020 - 34th AAAI Conference on Artificial Intelligence

Publication Date

January 1, 2020

Start / End Page

6829 / 6836