SpecNet: Spectral Domain Convolutional Neural Network

05/27/2019
by   Bochen Guan, et al.
0

The memory consumption of most Convolutional Neural Network (CNN) architectures grows rapidly with increasing depth of the network, which is a major constraint for efficient network training and inference on modern GPUs with yet limited memory. Several studies show that the feature maps (as generated after the convolutional layers) are the big bottleneck in this memory problem. Often, these feature maps mimic natural photographs in the sense that their energy is concentrated in the spectral domain. This paper proposes a Spectral Domain Convolutional Neural Network (SpecNet) that performs both the convolution and the activation operations in the spectral domain to achieve memory reduction. SpecNet exploits a configurable threshold to force small values in the feature maps to zero, allowing the feature maps to be stored sparsely. Since convolution in the spatial domain is equivalent to a dot product in the spectral domain, the multiplications only need to be performed on the non-zero entries of the (sparse) spectral domain feature maps. SpecNet also employs a special activation function that preserves the sparsity of the feature maps while effectively encouraging the convergence of the network. The performance of SpecNet is evaluated on three competitive object recognition benchmark tasks (MNIST, CIFAR-10, and SVHN), and compared with four state-of-the-art implementations (LeNet, AlexNet, VGG, and DenseNet). Overall, SpecNet is able to reduce memory consumption by about 60 loss of performance for all tested network architectures.

READ FULL TEXT
research
01/09/2022

ThreshNet: An Efficient DenseNet using Threshold Mechanism to Reduce Connections

With the continuous development of neural networks in computer vision ta...
research
07/21/2017

Memory-Efficient Implementation of DenseNets

The DenseNet architecture is highly computationally efficient as a resul...
research
05/24/2017

Deep Rotation Equivariant Network

Recently, learning equivariant representations has attracted considerabl...
research
10/05/2020

Mind the Pad – CNNs can Develop Blind Spots

We show how feature maps in convolutional networks are susceptible to sp...
research
08/08/2017

Prune the Convolutional Neural Networks with Sparse Shrink

Nowadays, it is still difficult to adapt Convolutional Neural Network (C...
research
12/06/2013

Understanding Deep Architectures using a Recursive Convolutional Network

A key challenge in designing convolutional network models is sizing them...
research
09/17/2019

Thanks for Nothing: Predicting Zero-Valued Activations with Lightweight Convolutional Neural Networks

Convolutional neural networks (CNNs) introduce state-of-the-art results ...

Please sign up or login with your details

Forgot password? Click here to reset