L^2NAS: Learning to Optimize Neural Architectures via Continuous-Action Reinforcement Learning

09/25/2021
by   Keith G. Mills, et al.
0

Neural architecture search (NAS) has achieved remarkable results in deep neural network design. Differentiable architecture search converts the search over discrete architectures into a hyperparameter optimization problem which can be solved by gradient descent. However, questions have been raised regarding the effectiveness and generalizability of gradient methods for solving non-convex architecture hyperparameter optimization problems. In this paper, we propose L^2NAS, which learns to intelligently optimize and update architecture hyperparameters via an actor neural network based on the distribution of high-performing architectures in the search history. We introduce a quantile-driven training procedure which efficiently trains L^2NAS in an actor-critic framework via continuous-action reinforcement learning. Experiments show that L^2NAS achieves state-of-the-art results on NAS-Bench-201 benchmark as well as DARTS search space and Once-for-All MobileNetV3 search space. We also show that search policies generated by L^2NAS are generalizable and transferable across different training datasets with minimal fine-tuning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/24/2018

SNAS: Stochastic Neural Architecture Search

We propose Stochastic Neural Architecture Search (SNAS), an economical e...
research
12/13/2018

IRLAS: Inverse Reinforcement Learning for Architecture Search

In this paper, we propose an inverse reinforcement learning method for a...
research
01/26/2019

A Practical Bandit Method with Advantages in Neural Network Tuning

Stochastic bandit algorithms can be used for challenging non-convex opti...
research
07/11/2020

An Asymptotically Optimal Multi-Armed Bandit Algorithm and Hyperparameter Optimization

The evaluation of hyperparameters, neural architectures, or data augment...
research
12/02/2019

GroSS: Group-Size Series Decomposition for Whole Search-Space Training

We present Group-size Series (GroSS) decomposition, a mathematical formu...
research
04/29/2021

Generalization Guarantees for Neural Architecture Search with Train-Validation Split

Neural Architecture Search (NAS) is a popular method for automatically d...
research
01/29/2020

Bayesian Neural Architecture Search using A Training-Free Performance Metric

Recurrent neural networks (RNNs) are a powerful approach for time series...

Please sign up or login with your details

Forgot password? Click here to reset