Deep Kernel Principal Component Analysis for Multi-level Feature Learning

02/22/2023
by   Francesco Tonin, et al.
0

Principal Component Analysis (PCA) and its nonlinear extension Kernel PCA (KPCA) are widely used across science and industry for data analysis and dimensionality reduction. Modern deep learning tools have achieved great empirical success, but a framework for deep principal component analysis is still lacking. Here we develop a deep kernel PCA methodology (DKPCA) to extract multiple levels of the most informative components of the data. Our scheme can effectively identify new hierarchical variables, called deep principal components, capturing the main characteristics of high-dimensional data through a simple and interpretable numerical optimization. We couple the principal components of multiple KPCA levels, theoretically showing that DKPCA creates both forward and backward dependency across levels, which has not been explored in kernel methods and yet is crucial to extract more informative features. Various experimental evaluations on multiple data types show that DKPCA finds more efficient and disentangled representations with higher explained variance in fewer principal components, compared to the shallow KPCA. We demonstrate that our method allows for effective hierarchical data exploration, with the ability to separate the key generative factors of the input data both for large datasets and when few training samples are available. Overall, DKPCA can facilitate the extraction of useful patterns from high-dimensional data by learning more informative features organized in different levels, giving diversified aspects to explore the variation factors in the data, while maintaining a simple mathematical formulation.

READ FULL TEXT

page 12

page 13

research
01/07/2020

A kernel Principal Component Analysis (kPCA) digest with a new backward mapping (pre-image reconstruction) strategy

Methodologies for multidimensionality reduction aim at discovering low-d...
research
11/25/2020

Unsupervised learning of disentangled representations in deep restricted kernel machines with orthogonality constraints

We introduce Constr-DRKM, a deep kernel method for the unsupervised lear...
research
10/07/2019

Push it to the Limit: Discover Edge-Cases in Image Data with Autoencoders

In this paper, we focus on the problem of identifying semantic factors o...
research
06/27/2021

Interpretable Network Representation Learning with Principal Component Analysis

We consider the problem of interpretable network representation learning...
research
03/17/2022

Dimensionality Reduction and Wasserstein Stability for Kernel Regression

In a high-dimensional regression framework, we study consequences of the...
research
07/19/2018

The Deep Kernelized Autoencoder

Autoencoders learn data representations (codes) in such a way that the i...
research
08/23/2022

Neural PCA for Flow-Based Representation Learning

Of particular interest is to discover useful representations solely from...

Please sign up or login with your details

Forgot password? Click here to reset