Deep Networks from the Principle of Rate Reduction

10/27/2020
by   Kwan Ho Ryan Chan, et al.
16

This work attempts to interpret modern deep (convolutional) networks from the principles of rate reduction and (shift) invariant classification. We show that the basic iterative gradient ascent scheme for optimizing the rate reduction of learned features naturally leads to a multi-layer deep network, one iteration per layer. The layered architectures, linear and nonlinear operators, and even parameters of the network are all explicitly constructed layer-by-layer in a forward propagation fashion by emulating the gradient scheme. All components of this "white box" network have precise optimization, statistical, and geometric interpretation. This principled framework also reveals and justifies the role of multi-channel lifting and sparse coding in early stage of deep networks. Moreover, all linear operators of the so-derived network naturally become multi-channel convolutions when we enforce classification to be rigorously shift-invariant. The derivation also indicates that such a convolutional network is significantly more efficient to construct and learn in the spectral domain. Our preliminary simulations and experiments indicate that so constructed deep network can already learn a good discriminative representation even without any back propagation training.

READ FULL TEXT

page 13

page 30

page 33

page 34

research
05/21/2021

ReduNet: A White-box Deep Network from the Principle of Maximizing Rate Reduction

This work attempts to provide a plausible theoretical framework that aim...
research
06/01/2023

White-Box Transformers via Sparse Rate Reduction

In this paper, we contend that the objective of representation learning ...
research
10/21/2019

Hyperspectral Image Classification Based on Adaptive Sparse Deep Network

Sparse model is widely used in hyperspectral image classification.Howeve...
research
04/25/2019

Making Convolutional Networks Shift-Invariant Again

Modern convolutional networks are not shift-invariant, as small input sh...
research
03/07/2022

Singular Value Perturbation and Deep Network Optimization

We develop new theoretical results on matrix perturbation to shed light ...
research
05/21/2019

Geometry of Deep Convolutional Networks

We give a formal procedure for computing preimages of convolutional netw...
research
11/16/2017

A Forward-Backward Approach for Visualizing Information Flow in Deep Networks

We introduce a new, systematic framework for visualizing information flo...

Please sign up or login with your details

Forgot password? Click here to reset