Fundamental Limits and Tradeoffs in Invariant Representation Learning

12/19/2020
by   Han Zhao, et al.
0

Many machine learning applications involve learning representations that achieve two competing goals: To maximize information or accuracy with respect to a subset of features (e.g. for prediction) while simultaneously maximizing invariance or independence with respect to another, potentially overlapping, subset of features (e.g. for fairness, privacy, etc). Typical examples include privacy-preserving learning, domain adaptation, and algorithmic fairness, just to name a few. In fact, all of the above problems admit a common minimax game-theoretic formulation, whose equilibrium represents a fundamental tradeoff between accuracy and invariance. Despite its abundant applications in the aforementioned domains, theoretical understanding on the limits and tradeoffs of invariant representations is severely lacking. In this paper, we provide an information-theoretic analysis of this general and important problem under both classification and regression settings. In both cases, we analyze the inherent tradeoffs between accuracy and invariance by providing a geometric characterization of the feasible region in the information plane, where we connect the geometric properties of this feasible region to the fundamental limitations of the tradeoff problem. In the regression setting, we also derive a tight lower bound on the Lagrangian objective that quantifies the tradeoff between accuracy and invariance. This lower bound leads to a better understanding of the tradeoff via the spectral properties of the joint distribution. In both cases, our results shed new light on this fundamental problem by providing insights on the interplay between accuracy and invariance. These results deepen our understanding of this fundamental problem and may be useful in guiding the design of adversarial representation learning algorithms.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/27/2019

On Learning Invariant Representation for Domain Adaptation

Due to the ability of deep neural nets to learn rich representations, re...
research
06/19/2019

Inherent Tradeoffs in Learning Fair Representation

With the prevalence of machine learning in high-stakes applications, esp...
research
01/23/2016

Minimax Lower Bounds for Linear Independence Testing

Linear independence testing is a fundamental information-theoretic and s...
research
05/31/2017

Controllable Invariance through Adversarial Feature Learning

Learning meaningful representations that maintain the content necessary ...
research
06/16/2021

Costs and Benefits of Wasserstein Fair Regression

Real-world applications of machine learning tools in high-stakes domains...
research
10/17/2019

An Information-Theoretic Perspective on the Relationship Between Fairness and Accuracy

Our goal is to understand the so-called trade-off between fairness and a...
research
03/01/2018

PIP Distance: A Unitary-invariant Metric for Understanding Functionality and Dimensionality of Vector Embeddings

In this paper, we present a theoretical framework for understanding vect...

Please sign up or login with your details

Forgot password? Click here to reset