ReduNet: A White-box Deep Network from the Principle of Maximizing Rate Reduction

05/21/2021
by   Kwan Ho Ryan Chan, et al.
16

This work attempts to provide a plausible theoretical framework that aims to interpret modern deep (convolutional) networks from the principles of data compression and discriminative representation. We argue that for high-dimensional multi-class data, the optimal linear discriminative representation maximizes the coding rate difference between the whole dataset and the average of all the subsets. We show that the basic iterative gradient ascent scheme for optimizing the rate reduction objective naturally leads to a multi-layer deep network, named ReduNet, which shares common characteristics of modern deep networks. The deep layered architectures, linear and nonlinear operators, and even parameters of the network are all explicitly constructed layer-by-layer via forward propagation, although they are amenable to fine-tuning via back propagation. All components of so-obtained “white-box” network have precise optimization, statistical, and geometric interpretation. Moreover, all linear operators of the so-derived network naturally become multi-channel convolutions when we enforce classification to be rigorously shift-invariant. The derivation in the invariant setting suggests a trade-off between sparsity and invariance, and also indicates that such a deep convolution network is significantly more efficient to construct and learn in the spectral domain. Our preliminary simulations and experiments clearly verify the effectiveness of both the rate reduction objective and the associated ReduNet. All code and data are available at https://github.com/Ma-Lab-Berkeley.

READ FULL TEXT
research
10/27/2020

Deep Networks from the Principle of Rate Reduction

This work attempts to interpret modern deep (convolutional) networks fro...
research
06/01/2023

White-Box Transformers via Sparse Rate Reduction

In this paper, we contend that the objective of representation learning ...
research
04/25/2019

Making Convolutional Networks Shift-Invariant Again

Modern convolutional networks are not shift-invariant, as small input sh...
research
11/22/2019

Bounding Singular Values of Convolution Layers

In deep neural networks, the spectral norm of the Jacobian of a layer bo...
research
11/30/2020

Incremental Learning via Rate Reduction

Current deep learning architectures suffer from catastrophic forgetting,...
research
10/16/2019

On the Global Optima of Kernelized Adversarial Representation Learning

Adversarial representation learning is a promising paradigm for obtainin...
research
11/30/2021

Playing Ping Pong with Light: Directional Emission of White Light

Over the last decades, light-emitting diodes (LED) have replaced common ...

Please sign up or login with your details

Forgot password? Click here to reset