Neural Architecture Search Using Stable Rank of Convolutional Layers

09/19/2020
by   Kengo Machida, et al.
18

In Neural Architecture Search (NAS), Differentiable ARchiTecture Search (DARTS) has recently attracted much attention due to its high efficiency. It defines an over-parameterized network with mixed edges each of which represents all operator candidates, and jointly optimizes the weights of the network and its architecture in an alternating way. However, this process prefers a model whose weights converge faster than the others, and such a model with fastest convergence often leads to overfitting. Accordingly the resulting model cannot always be well-generalized. To overcome this problem, we propose Minimum Stable Rank DARTS (MSR-DARTS), which aims to find a model with the best generalization error by replacing the architecture optimization with the selection process using the minimum stable rank criterion. Specifically, a convolution operator is represented by a matrix and our method chooses the one whose stable rank is the smallest. We evaluate MSR-DARTS on CIFAR-10 and ImageNet dataset. It achieves an error rate of 2.92 on CIFAR-10, and a top-1 error rate of 24.0 directly optimizes an ImageNet model with only 2.6 GPU days while it is often impractical for existing NAS methods to directly optimize a large model such as ImageNet models and hence a proxy dataset such as CIFAR-10 is often utilized.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/15/2020

Differentiable Neural Architecture Transformation for Reproducible Architecture Improvement

Recently, Neural Architecture Search (NAS) methods are introduced and sh...
research
11/24/2019

Exploiting Operation Importance for Differentiable Neural Architecture Search

Recently, differentiable neural architecture search methods significantl...
research
11/30/2020

Inter-layer Transition in Neural Architecture Search

Differential Neural Architecture Search (NAS) methods represent the netw...
research
10/16/2020

G-DARTS-A: Groups of Channel Parallel Sampling with Attention

Differentiable Architecture Search (DARTS) provides a baseline for searc...
research
02/19/2020

Knapsack Pruning with Inner Distillation

Neural network pruning reduces the computational cost of an over-paramet...
research
07/31/2020

Neural Architecture Search as Sparse Supernet

This paper aims at enlarging the problem of Neural Architecture Search f...
research
05/13/2021

Neighborhood-Aware Neural Architecture Search

Existing neural architecture search (NAS) methods often return an archit...

Please sign up or login with your details

Forgot password? Click here to reset