Towards Automatic Neural Architecture Search within General Super-Networks

05/25/2023
by   Tianyi Chen, et al.
0

Existing neural architecture search (NAS) methods typically rely on pre-specified super deep neural networks (super-networks) with handcrafted search spaces beforehand. Such requirements make it challenging to extend them onto general scenarios without significant human expertise and manual intervention. To overcome the limitations, we propose the third generation of Only-Train-Once (OTOv3). OTOv3 is perhaps the first automated system that trains general super-networks and produces high-performing sub-networks in the one shot manner without pretraining and fine-tuning. Technologically, OTOv3 delivers three noticeable contributions to minimize human efforts: (i) automatic search space construction for general super-networks; (ii) a Hierarchical Half-Space Projected Gradient (H2SPG) that leverages the dependency graph to ensure the network validity during optimization and reliably produces a solution with both high performance and hierarchical group sparsity; and (iii) automatic sub-network construction based on the super-network and the H2SPG solution. Numerically, we demonstrate the effectiveness of OTOv3 on a variety of super-networks, including RegNet, StackedUnets, SuperResNet, and DARTS, over benchmark datasets such as CIFAR10, Fashion-MNIST, ImageNet, STL-10, and SVNH. The sub-networks computed by OTOv3 achieve competitive even superior performance compared to the super-networks and other state-of-the-arts. The library will be released at https://github.com/tianyic/only_train_once.

READ FULL TEXT
research
06/11/2020

Few-shot Neural Architecture Search

To improve the search efficiency for Neural Architecture Search (NAS), O...
research
03/13/2023

OTOV2: Automatic, Generic, User-Friendly

The existing model compression methods via structured pruning typically ...
research
07/03/2023

Neural Architecture Transfer 2: A Paradigm for Improving Efficiency in Multi-Objective Neural Architecture Search

Deep learning is increasingly impacting various aspects of contemporary ...
research
12/20/2021

Enabling NAS with Automated Super-Network Generation

Recent Neural Architecture Search (NAS) solutions have produced impressi...
research
04/23/2021

Inter-choice dependent super-network weights

The automatic design of architectures for neural networks, Neural Archit...
research
02/25/2022

A Hardware-Aware System for Accelerating Deep Neural Network Optimization

Recent advances in Neural Architecture Search (NAS) which extract specia...
research
04/10/2021

Group Equivariant Neural Architecture Search via Group Decomposition and Reinforcement Learning

Recent works show that including group equivariance as an inductive bias...

Please sign up or login with your details

Forgot password? Click here to reset