Seesaw-Net: Convolution Neural Network With Uneven Group Convolution

05/09/2019
by   Jintao Zhang, et al.
0

In this paper, we are interested in boosting the representation capability of convolution neural networks which utilizing the inverted residual structure. Based on the success of Inverted Residual structure[Sandler et al. 2018] and Interleaved Low-Rank Group Convolutions[Sun et al. 2018], we rethink this two pattern of neural network structure, rather than NAS(Neural architecture search) method[Zoph and Le 2017; Pham et al. 2018; Liu et al. 2018b], we introduce uneven point-wise group convolution, which provide a novel search space for designing basic blocks to obtain better trade-off between representation capability and computational cost. Meanwhile, we propose two novel information flow patterns that will enable cross-group information flow for multiple group convolution layers with and without any channel permute/shuffle operation. Dense experiments on image classification task show that our proposed model, named Seesaw-Net, achieves state-of-the-art(SOTA) performance with limited computation and memory cost. Our code will be open-source and available together with pre-trained models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/01/2021

Training BatchNorm Only in Neural Architecture Search and Beyond

This work investigates the usage of batch normalization in neural archit...
research
09/22/2022

Equivariant Transduction through Invariant Alignment

The ability to generalize compositionally is key to understanding the po...
research
05/13/2020

Binarizing MobileNet via Evolution-based Searching

Binary Neural Networks (BNNs), known to be one among the effectively com...
research
07/31/2018

MnasNet: Platform-Aware Neural Architecture Search for Mobile

Designing convolutional neural networks (CNN) models for mobile devices ...
research
01/05/2022

Neural Architecture Search for Inversion

Over the year, people have been using deep learning to tackle inversion ...
research
08/17/2015

Molding CNNs for text: non-linear, non-consecutive convolutions

The success of deep learning often derives from well-chosen operational ...
research
10/07/2019

Energy-Aware Neural Architecture Optimization with Fast Splitting Steepest Descent

Designing energy-efficient networks is of critical importance for enabli...

Please sign up or login with your details

Forgot password? Click here to reset