Semi-Cyclic Stochastic Gradient Descent

04/23/2019
by   Hubert Eichner, et al.
0

We consider convex SGD updates with a block-cyclic structure, i.e. where each cycle consists of a small number of blocks, each with many samples from a possibly different, block-specific, distribution. This situation arises, e.g., in Federated Learning where the mobile devices available for updates at different times during the day have different characteristics. We show that such block-cyclic structure can significantly deteriorate the performance of SGD, but propose a simple approach that allows prediction with the same performance guarantees as for i.i.d., non-cyclic, sampling.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset