Differentiable Neural Architecture Learning for Efficient Neural Network Design

03/03/2021
by   Qingbei Guo, et al.
0

Automated neural network design has received ever-increasing attention with the evolution of deep convolutional neural networks (CNNs), especially involving their deployment on embedded and mobile platforms. One of the biggest problems that neural architecture search (NAS) confronts is that a large number of candidate neural architectures are required to train, using, for instance, reinforcement learning and evolutionary optimisation algorithms, at a vast computation cost. Even recent differentiable neural architecture search (DNAS) samples a small number of candidate neural architectures based on the probability distribution of learned architecture parameters to select the final neural architecture. To address this computational complexity issue, we introduce a novel architecture parameterisation based on scaled sigmoid function, and propose a general Differentiable Neural Architecture Learning (DNAL) method to optimize the neural architecture without the need to evaluate candidate neural networks. Specifically, for stochastic supernets as well as conventional CNNs, we build a new channel-wise module layer with the architecture components controlled by a scaled sigmoid function. We train these neural network models from scratch. The network optimization is decoupled into the weight optimization and the architecture optimization. We address the non-convex optimization problem of neural architecture by the continuous scaled sigmoid method with convergence guarantees. Extensive experiments demonstrate our DNAL method delivers superior performance in terms of neural architecture search cost. The optimal networks learned by DNAL surpass those produced by the state-of-the-art methods on the benchmark CIFAR-10 and ImageNet-1K dataset in accuracy, model size and computational complexity.

READ FULL TEXT

page 1

page 11

research
11/25/2019

Binarized Neural Architecture Search

Neural architecture search (NAS) can have a significant impact in comput...
research
08/16/2021

Probeable DARTS with Application to Computational Pathology

AI technology has made remarkable achievements in computational patholog...
research
06/05/2019

Butterfly Transform: An Efficient FFT Based Neural Architecture Design

In this paper, we introduce the Butterfly Transform (BFT), a light weigh...
research
09/11/2019

CARS: Continuous Evolution for Efficient Neural Architecture Search

Searching techniques in most of existing neural architecture search (NAS...
research
10/27/2018

Training Frankenstein's Creature to Stack: HyperTree Architecture Search

We propose HyperTrees for the low cost automatic design of multiple-inpu...
research
12/07/2017

Hungarian Layer: Logics Empowered Neural Architecture

Neural architecture is a purely numeric framework, which fits the data a...
research
08/08/2020

NASB: Neural Architecture Search for Binary Convolutional Neural Networks

Binary Convolutional Neural Networks (CNNs) have significantly reduced t...

Please sign up or login with your details

Forgot password? Click here to reset