Improved learning theory for kernel distribution regression with two-stage sampling

08/28/2023
by   François Bachoc, et al.
0

The distribution regression problem encompasses many important statistics and machine learning tasks, and arises in a large range of applications. Among various existing approaches to tackle this problem, kernel methods have become a method of choice. Indeed, kernel distribution regression is both computationally favorable, and supported by a recent learning theory. This theory also tackles the two-stage sampling setting, where only samples from the input distributions are available. In this paper, we improve the learning theory of kernel distribution regression. We address kernels based on Hilbertian embeddings, that encompass most, if not all, of the existing approaches. We introduce the novel near-unbiased condition on the Hilbertian embeddings, that enables us to provide new error bounds on the effect of the two-stage sampling, thanks to a new analysis. We show that this near-unbiased condition holds for three important classes of kernels, based on optimal transport and mean embedding. As a consequence, we strictly improve the existing convergence rates for these kernels. Our setting and results are illustrated by numerical experiments.

READ FULL TEXT
research
08/26/2022

Coefficient-based Regularized Distribution Regression

In this paper, we consider the coefficient-based regularized distributio...
research
02/08/2022

Distribution Regression with Sliced Wasserstein Kernels

The problem of learning functions over spaces of probabilities - or dist...
research
06/16/2020

Estimates on Learning Rates for Multi-Penalty Distribution Regression

This paper is concerned with functional learning by utilizing two-stage ...
research
02/07/2014

Two-stage Sampled Learning Theory on Distributions

We focus on the distribution regression problem: regressing to a real-va...
research
11/08/2014

Learning Theory for Distribution Regression

We focus on the distribution regression problem: regressing to vector-va...
research
10/15/2021

Kernel Minimum Divergence Portfolios

Portfolio optimization is a key challenge in finance with the aim of cre...
research
10/11/2019

A General Scoring Rule for Randomized Kernel Approximation with Application to Canonical Correlation Analysis

Random features has been widely used for kernel approximation in large-s...

Please sign up or login with your details

Forgot password? Click here to reset