What is the Best Feature Learning Procedure in Hierarchical Recognition Architectures?

06/05/2016
by   Kevin Jarrett, et al.
0

(This paper was written in November 2011 and never published. It is posted on arXiv.org in its original form in June 2016). Many recent object recognition systems have proposed using a two phase training procedure to learn sparse convolutional feature hierarchies: unsupervised pre-training followed by supervised fine-tuning. Recent results suggest that these methods provide little improvement over purely supervised systems when the appropriate nonlinearities are included. This paper presents an empirical exploration of the space of learning procedures for sparse convolutional networks to assess which method produces the best performance. In our study, we introduce an augmentation of the Predictive Sparse Decomposition method that includes a discriminative term (DPSD). We also introduce a new single phase supervised learning procedure that places an L1 penalty on the output state of each layer of the network. This forces the network to produce sparse codes without the expensive pre-training phase. Using DPSD with a new, complex predictor that incorporates lateral inhibition, combined with multi-scale feature pooling, and supervised refinement, the system achieves a 70.6% recognition rate on Caltech-101. With the addition of convolutional training, a 77% recognition was obtained on the CIfAR-10 dataset.

READ FULL TEXT

page 9

page 11

page 14

research
12/20/2014

An Analysis of Unsupervised Pre-training in Light of Recent Advances

Convolutional neural networks perform well on object recognition because...
research
06/27/2012

Large-Scale Feature Learning With Spike-and-Slab Sparse Coding

We consider the problem of object recognition with a large number of cla...
research
11/05/2019

MML: Maximal Multiverse Learning for Robust Fine-Tuning of Language Models

Recent state-of-the-art language models utilize a two-phase training pro...
research
02/01/2021

On Scaling Contrastive Representations for Low-Resource Speech Recognition

Recent advances in self-supervised learning through contrastive training...
research
01/19/2015

Statistical-mechanical analysis of pre-training and fine tuning in deep learning

In this paper, we present a statistical-mechanical analysis of deep lear...
research
12/03/2009

Behavior and performance of the deep belief networks on image classification

We apply deep belief networks of restricted Boltzmann machines to bags o...
research
03/13/2015

Sparse Code Formation with Linear Inhibition

Sparse code formation in the primary visual cortex (V1) has been inspira...

Please sign up or login with your details

Forgot password? Click here to reset