Path-Level Network Transformation for Efficient Architecture Search

06/07/2018
by   Han Cai, et al.
2

We introduce a new function-preserving transformation for efficient neural architecture search. This network transformation allows reusing previously trained networks and existing successful architectures that improves sample efficiency. We aim to address the limitation of current network transformation operations that can only perform layer-level architecture modifications, such as adding (pruning) filters or inserting (removing) a layer, which fails to change the topology of connection paths. Our proposed path-level transformation operations enable the meta-controller to modify the path topology of the given network while keeping the merits of reusing weights, and thus allow efficiently designing effective structures with complex path topologies like Inception models. We further propose a bidirectional tree-structured reinforcement learning meta-controller to explore a simple yet highly expressive tree-structured architecture space that can be viewed as a generalization of multi-branch architectures. We experimented on the image classification datasets with limited computational resources (about 200 GPU-hours), where we observed improved parameter efficiency and better test results (97.70 accuracy on CIFAR-10 with 14.3M parameters and 74.6 in the mobile setting), demonstrating the effectiveness and transferability of our designed architectures.

READ FULL TEXT

page 3

page 6

research
07/16/2017

Efficient Architecture Search by Network Transformation

Techniques for automatically designing deep neural network architectures...
research
10/13/2021

Improving the sample-efficiency of neural architecture search with reinforcement learning

Designing complex architectures has been an essential cogwheel in the re...
research
09/14/2019

Neural Architecture Search for Class-incremental Learning

In class-incremental learning, a model learns continuously from a sequen...
research
09/05/2019

Efficient Neural Architecture Transformation Searchin Channel-Level for Object Detection

Recently, Neural Architecture Search has achieved great success in large...
research
02/09/2018

Efficient Neural Architecture Search via Parameter Sharing

We propose Efficient Neural Architecture Search (ENAS), a fast and inexp...
research
12/03/2019

Multi-Criterion Evolutionary Design of Deep Convolutional Neural Networks

Convolutional neural networks (CNNs) are the backbones of deep learning ...
research
04/09/2019

SWNet: Small-World Neural Networks and Rapid Convergence

Training large and highly accurate deep learning (DL) models is computat...

Please sign up or login with your details

Forgot password? Click here to reset