DARTS+: Improved Differentiable Architecture Search with Early Stopping

09/13/2019
by   Hanwen Liang, et al.
0

Recently, there has been a growing interest in automating the process of neural architecture design, and the Differentiable Architecture Search (DARTS) method makes the process available within a few GPU days. In particular, a hyper-network called one-shot model is introduced, over which the architecture can be searched continuously with gradient descent. However, the performance of DARTS is often observed to collapse when the number of search epochs becomes large. Meanwhile, lots of "skip-connects" are found in the selected architectures. In this paper, we claim that the cause of the collapse is that there exist cooperation and competition in the bi-level optimization in DARTS, where the architecture parameters and model weights are updated alternatively. Therefore, we propose a simple and effective algorithm, named "DARTS+", to avoid the collapse and improve the original DARTS, by "early stopping" the search procedure when meeting a certain criterion. We demonstrate that the proposed early stopping criterion is effective in avoiding the collapse issue. We also conduct experiments on benchmark datasets and show the effectiveness of our DARTS+ algorithm, where DARTS+ achieves 2.32% test error on CIFAR10, 14.87% on CIFAR100, and 23.7% on ImageNet. We further remark that the idea of "early stopping" is implicitly included in some existing DARTS variants by manually setting a small number of search epochs, while we give an explicit criterion for "early stopping".

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/21/2021

Making Differentiable Architecture Search less local

Neural architecture search (NAS) is a recent methodology for automating ...
research
08/07/2023

Generalized Early Stopping in Evolutionary Direct Policy Search

Lengthy evaluation times are common in many optimization problems such a...
research
05/30/2019

Differentiable Neural Architecture Search via Proximal Iterations

Neural architecture search (NAS) recently attracts much research attenti...
research
08/25/2022

SONAR: Joint Architecture and System Optimization Search

There is a growing need to deploy machine learning for different tasks o...
research
09/16/2014

Compute Less to Get More: Using ORC to Improve Sparse Filtering

Sparse Filtering is a popular feature learning algorithm for image class...
research
05/07/2020

Noisy Differentiable Architecture Search

Simplicity is the ultimate sophistication. Differentiable Architecture S...
research
04/19/2022

Poisons that are learned faster are more effective

Imperceptible poisoning attacks on entire datasets have recently been to...

Please sign up or login with your details

Forgot password? Click here to reset