Adapting to Mixing Time in Stochastic Optimization with Markovian Data

02/09/2022
by   Ron Dorfman, et al.
0

We consider stochastic optimization problems where data is drawn from a Markov chain. Existing methods for this setting crucially rely on knowing the mixing time of the chain, which in real-world applications is usually unknown. We propose the first optimization method that does not require the knowledge of the mixing time, yet obtains the optimal asymptotic convergence rate when applied to convex problems. We further show that our approach can be extended to: (i) finding stationary points in non-convex optimization with Markovian data, and (ii) obtaining better dependence on the mixing time in temporal difference (TD) learning; in both cases, our method is completely oblivious to the mixing time. Our method relies on a novel combination of multi-level Monte Carlo (MLMC) gradient estimation together with an adaptive learning method.

READ FULL TEXT
research
05/27/2022

Constrained Langevin Algorithms with L-mixing External Random Variables

Langevin algorithms are gradient descent methods augmented with additive...
research
06/01/2019

Data-Pooling in Stochastic Optimization

Managing large-scale systems often involves simultaneously solving thous...
research
01/15/2018

Mixing Time on the Kagome Lattice

We consider tilings of a closed region of the Kagome lattice (partition ...
research
08/24/2017

Mixing time estimation in reversible Markov chains from a single sample path

The spectral gap γ of a finite, ergodic, and reversible Markov chain is ...
research
06/05/2021

Network Estimation by Mixing: Adaptivity and More

Networks analysis has been commonly used to study the interactions betwe...
research
06/25/2021

Bayesian Inference in High-Dimensional Time-Serieswith the Orthogonal Stochastic Linear Mixing Model

Many modern time-series datasets contain large numbers of output respons...

Please sign up or login with your details

Forgot password? Click here to reset