Neural architecture search as program transformation exploration

7Citations
Citations of this article
43Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Improving the performance of deep neural networks (DNNs) is important to both the compiler and neural architecture search (NAS) communities. Compilers apply program transformations in order to exploit hardware parallelism and memory hierarchy. However, legality concerns mean they fail to exploit the natural robustness of neural networks. In contrast, NAS techniques mutate networks by operations such as the grouping or bottlenecking of convolutions, exploiting the resilience of DNNs. In this work, we express such neural architecture operations as program transformations whose legality depends on a notion of representational capacity. This allows them to be combined with existing transformations into a unified optimization framework. This unification allows us to express existing NAS operations as combinations of simpler transformations. Crucially, it allows us to generate and explore new tensor convolutions. We prototyped the combined framework in TVM and were able to find optimizations across different DNNs, that significantly reduce inference time-over 3× in the majority of cases. Furthermore, our scheme dramatically reduces NAS search time.

Author supplied keywords

Cite

CITATION STYLE

APA

Turner, J., Crowley, E. J., & O’Boyle, M. F. P. (2021). Neural architecture search as program transformation exploration. In International Conference on Architectural Support for Programming Languages and Operating Systems - ASPLOS (pp. 915–927). Association for Computing Machinery. https://doi.org/10.1145/3445814.3446753

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free