Bootstrap Confidence Regions for Learned Feature Embeddings

02/01/2022
by   Kris Sankaran, et al.
0

Algorithmic feature learners provide high-dimensional vector representations for non-matrix structured signals, like images, audio, text, and graphs. Low-dimensional projections derived from these representations can be used to explore variation across collections of these data. However, it is not clear how to assess the uncertainty associated with these projections. We adapt methods developed for bootstrapping principal components analysis to the setting where features are learned from non-matrix data. We empirically compare the derived confidence regions in simulations, varying factors that influence both feature learning and the bootstrap. Approaches are illustrated on spatial proteomic data. Code, data, and trained models are released as an R compendium.

READ FULL TEXT

page 10

page 12

page 14

page 17

page 22

research
04/15/2021

Rates of Bootstrap Approximation for Eigenvalues in High-Dimensional PCA

In the context of principal components analysis (PCA), the bootstrap is ...
research
12/01/2021

AR-sieve Bootstrap for High-dimensional Time Series

This paper proposes a new AR-sieve bootstrap approach on high-dimensiona...
research
10/23/2022

Optimal Discriminant Analysis in High-Dimensional Latent Factor Models

In high-dimensional classification problems, a commonly used approach is...
research
06/28/2017

Asymptotic Confidence Regions for High-dimensional Structured Sparsity

In the setting of high-dimensional linear regression models, we propose ...
research
05/19/2021

A Note on High-Dimensional Confidence Regions

Recent advances in statistics introduced versions of the central limit t...
research
04/15/2022

BYOL for Audio: Exploring Pre-trained General-purpose Audio Representations

Pre-trained models are essential as feature extractors in modern machine...

Please sign up or login with your details

Forgot password? Click here to reset