Nonsparse learning with latent variables

10/07/2017
by   Zemin Zheng, et al.
0

As a popular tool for producing meaningful and interpretable models, large-scale sparse learning works efficiently when the underlying structures are indeed or close to sparse. However, naively applying the existing regularization methods can result in misleading outcomes due to model misspecification. In particular, the direct sparsity assumption on coefficient vectors has been questioned in real applications. Therefore, we consider nonsparse learning with the conditional sparsity structure that the coefficient vector becomes sparse after taking out the impacts of certain unobservable latent variables. A new methodology of nonsparse learning with latent variables (NSL) is proposed to simultaneously recover the significant observable predictors and latent factors as well as their effects. We explore a common latent family incorporating population principal components and derive the convergence rates of both sample principal components and their score vectors that hold for a wide class of distributions. With the properly estimated latent variables, properties including model selection consistency and oracle inequalities under various prediction and estimation losses are established for the proposed methodology. Our new methodology and results are evidenced by simulation and real data examples.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/10/2012

Asymptotic Accuracy of Distribution-Based Estimation for Latent Variables

Hierarchical statistical models are widely employed in information scien...
research
07/31/2021

Estimation and visualization of treatment effects for multiple outcomes

We consider a randomized controlled trial between two groups. The object...
research
01/25/2021

Identifying Interpretable Discrete Latent Structures from Discrete Data

High dimensional categorical data are routinely collected in biomedical ...
research
05/09/2017

SILVar: Single Index Latent Variable Models

A semi-parametric, non-linear regression model in the presence of latent...
research
01/02/2022

Factor tree copula models for item response data

Factor copula models for item response data are more interpretable and f...
research
06/05/2021

Principal Bit Analysis: Autoencoding with Schur-Concave Loss

We consider a linear autoencoder in which the latent variables are quant...
research
07/18/2021

Impedance-based Capacity Estimation for Lithium-Ion Batteries Using Generative Adversarial Network

This paper proposes a fully unsupervised methodology for the reliable ex...

Please sign up or login with your details

Forgot password? Click here to reset