Learning from weakly dependent data under Dobrushin's condition

06/21/2019
by   Yuval Dagan, et al.
0

Statistical learning theory has largely focused on learning and generalization given independent and identically distributed (i.i.d.) samples. Motivated by applications involving time-series data, there has been a growing literature on learning and generalization in settings where data is sampled from an ergodic process. This work has also developed complexity measures, which appropriately extend the notion of Rademacher complexity to bound the generalization error and learning rates of hypothesis classes in this setting. Rather than time-series data, our work is motivated by settings where data is sampled on a network or a spatial domain, and thus do not fit well within the framework of prior work. We provide learning and generalization bounds for data that are complexly dependent, yet their distribution satisfies the standard Dobrushin's condition. Indeed, we show that the standard complexity measures of Gaussian and Rademacher complexities and VC dimension are sufficient measures of complexity for the purposes of bounding the generalization error and learning rates of hypothesis classes in our setting. Moreover, our generalization bounds only degrade by constant factors compared to their i.i.d. analogs, and our learnability bounds degrade by log factors in the size of the training set.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/09/2019

Hypothesis Set Stability and Generalization

We present an extensive study of generalization for data-dependent hypot...
research
07/04/2023

Generalization Guarantees via Algorithm-dependent Rademacher Complexity

Algorithm- and data-dependent generalization bounds are required to expl...
research
06/03/2011

Rademacher complexity of stationary sequences

We show how to control the generalization error of time series models wh...
research
05/30/2014

Generalization Bounds for Learning with Linear, Polygonal, Quadratic and Conic Side Knowledge

In this paper, we consider a supervised learning setting where side know...
research
03/21/2023

Uniform Risk Bounds for Learning with Dependent Data Sequences

This paper extends standard results from learning theory with independen...
research
02/14/2016

Generalization Properties of Learning with Random Features

We study the generalization properties of ridge regression with random f...
research
05/12/2015

Permutational Rademacher Complexity: a New Complexity Measure for Transductive Learning

Transductive learning considers situations when a learner observes m lab...

Please sign up or login with your details

Forgot password? Click here to reset