DeepMAD: Mathematical Architecture Design for Deep Convolutional Neural Network

03/05/2023
by   Xuan Shen, et al.
0

The rapid advances in Vision Transformer (ViT) refresh the state-of-the-art performances in various vision tasks, overshadowing the conventional CNN-based models. This ignites a few recent striking-back research in the CNN world showing that pure CNN models can achieve as good performance as ViT models when carefully tuned. While encouraging, designing such high-performance CNN models is challenging, requiring non-trivial prior knowledge of network design. To this end, a novel framework termed Mathematical Architecture Design for Deep CNN (DeepMAD) is proposed to design high-performance CNN models in a principled way. In DeepMAD, a CNN network is modeled as an information processing system whose expressiveness and effectiveness can be analytically formulated by their structural parameters. Then a constrained mathematical programming (MP) problem is proposed to optimize these structural parameters. The MP problem can be easily solved by off-the-shelf MP solvers on CPUs with a small memory footprint. In addition, DeepMAD is a pure mathematical framework: no GPU or training data is required during network design. The superiority of DeepMAD is validated on multiple large-scale computer vision benchmark datasets. Notably on ImageNet-1k, only using conventional convolutional layers, DeepMAD achieves 0.7 0.8

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/26/2018

Effects of Degradations on Deep Neural Network Architectures

Deep convolutional neural networks (CNNs) have achieved many state-of-th...
research
10/03/2022

Introducing Vision Transformer for Alzheimer's Disease classification task with 3D input

Many high-performance classification models utilize complex CNN-based ar...
research
06/28/2021

Rethinking Token-Mixing MLP for MLP-based Vision Backbone

In the past decade, we have witnessed rapid progress in the machine visi...
research
01/29/2015

On Vectorization of Deep Convolutional Neural Networks for Vision Tasks

We recently have witnessed many ground-breaking results in machine learn...
research
06/02/2022

EfficientFormer: Vision Transformers at MobileNet Speed

Vision Transformers (ViT) have shown rapid progress in computer vision t...
research
11/14/2018

Interpretable deep learning for guided structure-property explorations in photovoltaics

The performance of an organic photovoltaic device is intricately connect...
research
08/10/2021

Exploiting Features with Split-and-Share Module

Deep convolutional neural networks (CNNs) have shown state-of-the-art pe...

Please sign up or login with your details

Forgot password? Click here to reset