BlockQNN: Efficient Block-wise Neural Network Architecture Generation

08/16/2018
by   Zhao Zhong, et al.
4

Convolutional neural networks have gained a remarkable success in computer vision. However, most usable network architectures are hand-crafted and usually require expertise and elaborate design. In this paper, we provide a block-wise network generation pipeline called BlockQNN which automatically builds high-performance networks using the Q-Learning paradigm with epsilon-greedy exploration strategy. The optimal network block is constructed by the learning agent which is trained to choose component layers sequentially. We stack the block to construct the whole auto-generated network. To accelerate the generation process, we also propose a distributed asynchronous framework and an early stop strategy. The block-wise generation brings unique advantages: (1) it yields state-of-the-art results in comparison to the hand-crafted networks on image classification, particularly, the best network generated by BlockQNN achieves 2.35 of the search space in designing networks, spending only 3 days with 32 GPUs. A faster version can yield a comparable result with only 1 GPU in 20 hours. (3) it has strong generalizability in that the network built on CIFAR also performs well on the larger-scale dataset. The best network achieves very competitive accuracy of 82.0

READ FULL TEXT
research
10/31/2018

Structure Learning of Deep Neural Networks with Q-Learning

Recently, with convolutional neural networks gaining significant achieve...
research
07/21/2017

Learning Transferable Architectures for Scalable Image Recognition

Developing neural network image classification models often requires sig...
research
10/10/2018

Automatic Configuration of Deep Neural Networks with EGO

Designing the architecture for an artificial neural network is a cumbers...
research
12/14/2018

Rethinking Layer-wise Feature Amounts in Convolutional Neural Network Architectures

We characterize convolutional neural networks with respect to the relati...
research
08/26/2021

Scalable and Modular Robustness Analysis of Deep Neural Networks

As neural networks are trained to be deeper and larger, the scalability ...
research
09/26/2019

Exascale Deep Learning to Accelerate Cancer Research

Deep learning, through the use of neural networks, has demonstrated rema...
research
07/22/2019

MemNet: Memory-Efficiency Guided Neural Architecture Search with Augment-Trim learning

Recent studies on automatic neural architectures search have demonstrate...

Please sign up or login with your details

Forgot password? Click here to reset