Fitting the Search Space of Weight-sharing NAS with Graph Convolutional Networks

7Citations
Citations of this article
39Readers
Mendeley users who have this article in their library.

Abstract

Neural architecture search has attracted wide attentions in both academia and industry. To accelerate it, researchers proposed weight-sharing methods which first train a super-network to reuse computation among different operators, from which exponentially many sub-networks can be sampled and efficiently evaluated. These methods enjoy great advantages in terms of computational costs, but the sampled sub-networks are not guaranteed to be estimated precisely unless an individual training process is taken. This paper attributes such inaccuracy to the inevitable mismatch between assembled network layers, so that there is a random error term added to each estimation. We alleviate this issue by training a graph convolutional network to fit the performance of sampled sub-networks so that the impact of random errors becomes minimal. With this strategy, we achieve a higher rank correlation coefficient in the selected set of candidates, which consequently leads to better performance of the final architecture. In addition, our approach also enjoys the flexibility of being used under different hardware constraints, since the graph convolutional network has provided an efficient lookup table of the performance of architectures in the entire search space.

Cite

CITATION STYLE

APA

Chen, X., Xie, L., Wu, J., Wei, L., Xu, Y., & Tian, Q. (2021). Fitting the Search Space of Weight-sharing NAS with Graph Convolutional Networks. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 8B, pp. 7064–7072). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i8.16869

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free