Improved Training of Sparse Coding Variational Autoencoder via Weight Normalization

01/23/2021
by   Linxing Preston Jiang, et al.
7

Learning a generative model of visual information with sparse and compositional features has been a challenge for both theoretical neuroscience and machine learning communities. Sparse coding models have achieved great success in explaining the receptive fields of mammalian primary visual cortex with sparsely activated latent representation. In this paper, we focus on a recently proposed model, sparse coding variational autoencoder (SVAE) (Barello et al., 2018), and show that the end-to-end training scheme of SVAE leads to a large group of decoding filters not fully optimized with noise-like receptive fields. We propose a few heuristics to improve the training of SVAE and show that a unit L_2 norm constraint on the decoder is critical to produce sparse coding filters. Such normalization can be considered as local lateral inhibition in the cortex. We verify this claim empirically on both natural image patches and MNIST dataset and show that projection of the filters onto unit norm drastically increases the number of active filters. Our results highlight the importance of weight normalization for learning sparse representation from data and suggest a new way of reducing the number of inactive latent components in VAE learning.

READ FULL TEXT

page 4

page 5

page 6

page 7

page 9

page 10

page 11

page 12

research
02/22/2023

Sparse, Geometric Autoencoder Models of V1

The classical sparse coding model represents visual stimuli as a linear ...
research
06/01/2022

Top-down inference in an early visual cortex inspired hierarchical Variational Autoencoder

Interpreting computations in the visual cortex as learning and inference...
research
03/29/2023

SC-VAE: Sparse Coding-based Variational Autoencoder

Learning rich data representations from unlabeled data is a key challeng...
research
02/26/2016

A Single Model Explains both Visual and Auditory Precortical Coding

Precortical neural systems encode information collected by the senses, b...
research
11/11/2018

Integrating Multiple Receptive Fields through Grouped Active Convolution

Convolutional networks have achieved great success in various vision tas...
research
07/30/2020

Quantitative Understanding of VAE by Interpreting ELBO as Rate Distortion Cost of Transform Coding

VAE (Variational autoencoder) estimates the posterior parameters (mean a...
research
09/30/2020

RG-Flow: A hierarchical and explainable flow model based on renormalization group and sparse prior

Flow-based generative models have become an important class of unsupervi...

Please sign up or login with your details

Forgot password? Click here to reset