Learning Theory for Distribution Regression

11/08/2014
by   Zoltan Szabo, et al.
0

We focus on the distribution regression problem: regressing to vector-valued outputs from probability measures. Many important machine learning and statistical tasks fit into this framework, including multi-instance learning and point estimation problems without analytical solution (such as hyperparameter or entropy estimation). Despite the large number of available heuristics in the literature, the inherent two-stage sampled nature of the problem makes the theoretical analysis quite challenging, since in practice only samples from sampled distributions are observable, and the estimates have to rely on similarities computed between sets of points. To the best of our knowledge, the only existing technique with consistency guarantees for distribution regression requires kernel density estimation as an intermediate step (which often performs poorly in practice), and the domain of the distributions to be compact Euclidean. In this paper, we study a simple, analytically computable, ridge regression-based alternative to distribution regression, where we embed the distributions to a reproducing kernel Hilbert space, and learn the regressor from the embeddings to the outputs. Our main contribution is to prove that this scheme is consistent in the two-stage sampled setup under mild conditions (on separable topological domains enriched with kernels): we present an exact computational-statistical efficiency trade-off analysis showing that our estimator is able to match the one-stage sampled minimax optimal rate [Caponnetto and De Vito, 2007; Steinwart et al., 2009]. This result answers a 17-year-old open question, establishing the consistency of the classical set kernel [Haussler, 1999; Gaertner et. al, 2002] in regression. We also cover consistency for more recent kernels on distributions, including those due to [Christmann and Steinwart, 2010].

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/07/2014

Two-stage Sampled Learning Theory on Distributions

We focus on the distribution regression problem: regressing to a real-va...
research
06/16/2020

Estimates on Learning Rates for Multi-Penalty Distribution Regression

This paper is concerned with functional learning by utilizing two-stage ...
research
08/26/2022

Coefficient-based Regularized Distribution Regression

In this paper, we consider the coefficient-based regularized distributio...
research
02/08/2022

Distribution Regression with Sliced Wasserstein Kernels

The problem of learning functions over spaces of probabilities - or dist...
research
08/28/2023

Improved learning theory for kernel distribution regression with two-stage sampling

The distribution regression problem encompasses many important statistic...
research
11/04/2014

Kernel Mean Estimation via Spectral Filtering

The problem of estimating the kernel mean in a reproducing kernel Hilber...
research
04/21/2021

Robust Kernel-based Distribution Regression

Regularization schemes for regression have been widely studied in learni...

Please sign up or login with your details

Forgot password? Click here to reset