Benign Autoencoders

10/02/2022
by   Semyon Malamud, et al.
0

The success of modern machine learning algorithms depends crucially on efficient data representation and compression through dimensionality reduction. This practice seemingly contradicts the conventional intuition suggesting that data processing always leads to information loss. We prove that this intuition is wrong. For any non-convex problem, there exists an optimal, benign auto-encoder (BAE) extracting a lower-dimensional data representation that is strictly beneficial: Compressing model inputs improves model performance. We prove that BAE projects data onto a manifold whose dimension is the compressibility dimension of the learning model. We develop and implement an efficient algorithm for computing BAE and show that BAE improves model performance in every dataset we consider. Furthermore, by compressing "malignant" data dimensions, BAE makes learning more stable and robust.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/29/2017

Dimensionality reduction methods for molecular simulations

Molecular simulations produce very high-dimensional data-sets with milli...
research
08/14/2019

Tensor-Train Parameterization for Ultra Dimensionality Reduction

Locality preserving projections (LPP) are a classical dimensionality red...
research
03/01/2018

Autoencoding topology

The problem of learning a manifold structure on a dataset is framed in t...
research
02/25/2016

Auto-JacoBin: Auto-encoder Jacobian Binary Hashing

Binary codes can be used to speed up nearest neighbor search tasks in la...
research
07/29/2013

Borel Isomorphic Dimensionality Reduction of Data and Supervised Learning

In this project we further investigate the idea of reducing the dimensio...
research
06/18/2020

The Dilemma Between Dimensionality Reduction and Adversarial Robustness

Recent work has shown the tremendous vulnerability to adversarial sample...

Please sign up or login with your details

Forgot password? Click here to reset