How far can we go without convolution: Improving fully-connected networks

11/09/2015
by   Zhouhan Lin, et al.
0

We propose ways to improve the performance of fully connected networks. We found that two approaches in particular have a strong effect on performance: linear bottleneck layers and unsupervised pre-training using autoencoders without hidden unit biases. We show how both approaches can be related to improving gradient flow and reducing sparsity in the network. We show that a fully connected network can yield approximately 70 the permutation-invariant CIFAR-10 task, which is much higher than the current state-of-the-art. By adding deformations to the training data, the fully connected network achieves 78 convolutional network.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/03/2020

Computational Separation Between Convolutional and Fully-Connected Networks

Convolutional neural networks (CNN) exhibit unmatched performance in a m...
research
06/08/2016

Convolution by Evolution: Differentiable Pattern Producing Networks

In this work we introduce a differentiable version of the Compositional ...
research
07/27/2020

Towards Learning Convolutions from Scratch

Convolution is one of the most essential components of architectures use...
research
11/06/2017

Characterizing Sparse Connectivity Patterns in Neural Networks

We propose a novel way of reducing the number of parameters in the stora...
research
10/16/2020

Why Are Convolutional Nets More Sample-Efficient than Fully-Connected Nets?

Convolutional neural networks often dominate fully-connected counterpart...
research
04/13/2019

Rethinking Classification and Localization in R-CNN

Modern R-CNN based detectors share the RoI feature extractor head for bo...
research
02/11/2020

Neural Rule Ensembles: Encoding Sparse Feature Interactions into Neural Networks

Artificial Neural Networks form the basis of very powerful learning meth...

Please sign up or login with your details

Forgot password? Click here to reset