Anytime Tail Averaging

02/13/2019
by   Nicolas Le Roux, et al.
0

Tail averaging consists in averaging the last examples in a stream. Common techniques either have a memory requirement which grows with the number of samples to average, are not available at every timestep or do not accomodate growing windows. We propose two techniques with a low constant memory cost that perform tail averaging with access to the average at every time step. We also show how one can improve the accuracy of that average at the cost of increased memory consumption.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/22/2019

Beating SGD Saturation with Tail-Averaging and Minibatching

While stochastic gradient descent (SGD) is one of the major workhorses i...
research
09/26/2022

Two-Tailed Averaging: Anytime Adaptive Once-in-a-while Optimal Iterate Averaging for Stochastic Optimization

Tail averaging improves on Polyak averaging's non-asymptotic behaviour b...
research
10/12/2022

Finite time analysis of temporal difference learning with linear function approximation: Tail averaging and regularisation

We study the finite-time behaviour of the popular temporal difference (T...
research
04/02/2019

Optimal designs for model averaging in non-nested models

In this paper we construct optimal designs for frequentist model averagi...
research
04/06/2023

PopulAtion Parameter Averaging (PAPA)

Ensemble methods combine the predictions of multiple models to improve p...
research
11/30/2022

Multi-level Parareal algorithm with Averaging

The present study is an extension of the work done in [16] and [10], whe...
research
06/07/2018

Re-evaluating evaluation

Progress in machine learning is measured by careful evaluation on proble...

Please sign up or login with your details

Forgot password? Click here to reset