Function Classes for Identifiable Nonlinear Independent Component Analysis

08/12/2022
by   Simon Buchholz, et al.
8

Unsupervised learning of latent variable models (LVMs) is widely used to represent data in machine learning. When such models reflect the ground truth factors and the mechanisms mapping them to observations, there is reason to expect that they allow generalization in downstream tasks. It is however well known that such identifiability guaranties are typically not achievable without putting constraints on the model class. This is notably the case for nonlinear Independent Component Analysis, in which the LVM maps statistically independent variables to observations via a deterministic nonlinear function. Several families of spurious solutions fitting perfectly the data, but that do not correspond to the ground truth factors can be constructed in generic settings. However, recent work suggests that constraining the function class of such models may promote identifiability. Specifically, function classes with constraints on their partial derivatives, gathered in the Jacobian matrix, have been proposed, such as orthogonal coordinate transformations (OCT), which impose orthogonality of the Jacobian columns. In the present work, we prove that a subclass of these transformations, conformal maps, is identifiable and provide novel theoretical results suggesting that OCTs have properties that prevent families of spurious solutions to spoil identifiability in a generic setting.

READ FULL TEXT

page 8

page 41

page 42

research
06/28/2023

Identifiability of Discretized Latent Coordinate Systems via Density Landmarks Detection

Disentanglement aims to recover meaningful latent ground-truth factors f...
research
05/26/2023

Causal Component Analysis

Independent Component Analysis (ICA) aims to recover independent latent ...
research
10/28/2021

Aggregation as Unsupervised Learning and its Evaluation

Regression uses supervised machine learning to find a model that combine...
research
02/14/2022

On Pitfalls of Identifiability in Unsupervised Learning. A Note on: "Desiderata for Representation Learning: A Causal Perspective"

Model identifiability is a desirable property in the context of unsuperv...
research
10/07/2021

Boxhead: A Dataset for Learning Hierarchical Representations

Disentanglement is hypothesized to be beneficial towards a number of dow...
research
06/14/2022

On Finite-Sample Identifiability of Contrastive Learning-Based Nonlinear Independent Component Analysis

Nonlinear independent component analysis (nICA) aims at recovering stati...
research
06/28/2022

Disentangling Embedding Spaces with Minimal Distributional Assumptions

Interest in understanding and factorizing learned embedding spaces is gr...

Please sign up or login with your details

Forgot password? Click here to reset