Neural Architecture Refinement: A Practical Way for Avoiding Overfitting in NAS

05/07/2019
by   Yang Jiang, et al.
0

Neural architecture search (NAS) is proposed to automate the architecture design process and attracts overwhelming interest from both academia and industry. However, it is confronted with overfitting issue due to the high-dimensional search space composed by operator selection and skip connection of each layer. This paper analyzes the overfitting issue from a novel perspective, which separates the primitives of search space into architecture-overfitting related and parameter-overfitting related elements. The operator of each layer, which mainly contributes to parameter-overfitting and is important for model acceleration, is selected as our optimization target based on state-of-the-art architecture, meanwhile skip which related to architecture-overfitting, is ignored. With the largely reduced search space, our proposed method is both quick to converge and practical to use in various tasks. Extensive experiments have demonstrated that the proposed method can achieve fascinated results, including classification, face recognition etc.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/28/2020

Angle-based Search Space Shrinking for Neural Architecture Search

In this work, we present a simple and general search space shrinking met...
research
12/23/2019

TextNAS: A Neural Architecture Search Space tailored for Text Representation

Learning text representation is crucial for text classification and othe...
research
03/19/2021

GNAS: A Generalized Neural Network Architecture Search Framework

In practice, the problems encountered in training NAS (Neural Architectu...
research
05/21/2019

Adaptive Stochastic Natural Gradient Method for One-Shot Neural Architecture Search

High sensitivity of neural architecture search (NAS) methods against the...
research
12/01/2021

Training BatchNorm Only in Neural Architecture Search and Beyond

This work investigates the usage of batch normalization in neural archit...
research
03/14/2023

Learning to Grow Artificial Hippocampi in Vision Transformers for Resilient Lifelong Learning

Lifelong learning without catastrophic forgetting (i.e., resiliency) pos...
research
02/20/2021

Towards Accurate and Compact Architectures via Neural Architecture Transformer

Designing effective architectures is one of the key factors behind the s...

Please sign up or login with your details

Forgot password? Click here to reset