ZARTS: On Zero-order Optimization for Neural Architecture Search

10/10/2021
by   Xiaoxing Wang, et al.
0

Differentiable architecture search (DARTS) has been a popular one-shot paradigm for NAS due to its high efficiency. It introduces trainable architecture parameters to represent the importance of candidate operations and proposes first/second-order approximation to estimate their gradients, making it possible to solve NAS by gradient descent algorithm. However, our in-depth empirical results show that the approximation will often distort the loss landscape, leading to the biased objective to optimize and in turn inaccurate gradient estimation for architecture parameters. This work turns to zero-order optimization and proposes a novel NAS scheme, called ZARTS, to search without enforcing the above approximation. Specifically, three representative zero-order optimization methods are introduced: RS, MGS, and GLD, among which MGS performs best by balancing the accuracy and speed. Moreover, we explore the connections between RS/MGS and gradient descent algorithm and show that our ZARTS can be seen as a robust gradient-free counterpart to DARTS. Extensive experiments on multiple datasets and search spaces show the remarkable performance of our method. In particular, results on 12 benchmarks verify the outstanding robustness of ZARTS, where the performance of DARTS collapses due to its known instability issue. Also, we search on the search space of DARTS to compare with peer methods, and our discovered architecture achieves 97.54 accuracy on CIFAR-10 and 75.7 state-of-the-art performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/22/2021

Differentiable Architecture Search Without Training Nor Labels: A Pruning Perspective

With leveraging the weight-sharing and continuous relaxation to enable g...
research
06/20/2022

Shapley-NAS: Discovering Operation Contribution for Neural Architecture Search

In this paper, we propose a Shapley value based method to evaluate opera...
research
05/30/2019

Differentiable Neural Architecture Search via Proximal Iterations

Neural architecture search (NAS) recently attracts much research attenti...
research
06/12/2023

Robustifying DARTS by Eliminating Information Bypass Leakage via Explicit Sparse Regularization

Differentiable architecture search (DARTS) is a promising end to end NAS...
research
10/25/2019

Stabilizing DARTS with Amended Gradient Estimation on Architectural Parameters

Differentiable neural architecture search has been a popular methodology...
research
08/22/2022

SVD-NAS: Coupling Low-Rank Approximation and Neural Architecture Search

The task of compressing pre-trained Deep Neural Networks has attracted w...
research
03/29/2022

Generalizing Few-Shot NAS with Gradient Matching

Efficient performance estimation of architectures drawn from large searc...

Please sign up or login with your details

Forgot password? Click here to reset