Efficient Neural Architecture Search via Parameters Sharing

02/09/2018
by   Hieu Pham, et al.
0

We propose Efficient Neural Architecture Search (ENAS), a fast and inexpensive approach for automatic model design. In ENAS, a controller learns to discover neural network architectures by searching for an optimal subgraph within a large computational graph. The controller is trained with policy gradient to select a subgraph that maximizes the expected reward on the validation set. Meanwhile the model corresponding to the selected subgraph is trained to minimize a canonical cross entropy loss. Thanks to parameter sharing between child models, ENAS is fast: it delivers strong empirical performances using much fewer GPU-hours than all existing automatic model design approaches, and notably, 1000x less expensive than standard Neural Architecture Search. On the Penn Treebank dataset, ENAS discovers a novel architecture that achieves a test perplexity of 55.8, establishing a new state-of-the-art among all methods without post-training processing. On the CIFAR-10 dataset, ENAS designs novel architectures that achieve a test error of 2.89 (Zoph et al., 2018), whose test error is 2.65

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/09/2018

Efficient Neural Architecture Search via Parameter Sharing

We propose Efficient Neural Architecture Search (ENAS), a fast and inexp...
research
07/16/2023

Neural Architecture Retrieval

With the increasing number of new neural architecture designs and substa...
research
01/05/2022

Neural Architecture Search for Inversion

Over the year, people have been using deep learning to tackle inversion ...
research
01/18/2020

Efficient Neural Architecture Search: A Broad Version

Efficient Neural Architecture Search (ENAS) achieves novel efficiency fo...
research
12/24/2019

BETANAS: BalancEd TrAining and selective drop for Neural Architecture Search

Automatic neural architecture search techniques are becoming increasingl...
research
03/04/2022

WPNAS: Neural Architecture Search by jointly using Weight Sharing and Predictor

Weight sharing based and predictor based methods are two major types of ...
research
03/31/2019

Understanding Neural Architecture Search Techniques

Automatic methods for generating state-of-the-art neural network archite...

Please sign up or login with your details

Forgot password? Click here to reset