Faster Algorithms for Large-scale Machine Learning using Simple Sampling Techniques

01/18/2018
by   Vinod Kumar Chauhan, et al.
0

Now a days, the major challenge in machine learning is the `Big Data' challenge. The big data problems due to large number of data points or large number of features in each data point, or both, the training of models have become very slow. The training time has two major components: Time to access the data and time to process the data. In this paper, we have proposed one possible solution to handle the big data problems in machine learning. The focus is on reducing the training time through reducing data access time by proposing systematic sampling and cyclic/sequential sampling to select mini-batches from the dataset. To prove the effectiveness of proposed sampling techniques, we have used Empirical Risk Minimization, which is commonly used machine learning problem, for strongly convex and smooth case. The problem has been solved using SAG, SAGA, SVRG, SAAG-II and MBSGD (Mini-batched SGD), each using two step determination techniques, namely, constant step size and backtracking line search method. Theoretical results prove the same convergence for systematic sampling, cyclic sampling and the widely used random sampling technique, in expectation. Experimental results with bench marked datasets prove the efficacy of the proposed sampling techniques.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/24/2018

SAAGs: Biased Stochastic Variance Reduction Methods

Stochastic optimization is one of the effective approach to deal with th...
research
04/19/2021

Random Reshuffling with Variance Reduction: New Analysis and Better Rates

Virtually all state-of-the-art methods for training supervised machine l...
research
07/25/2022

On the convergence and sampling of randomized primal-dual algorithms and their application to parallel MRI reconstruction

The Stochastic Primal-Dual Hybrid Gradient or SPDHG is an algorithm prop...
research
11/13/2016

Accelerated Variance Reduced Block Coordinate Descent

Algorithms with fast convergence, small number of data access, and low p...
research
03/10/2022

Differentially Private Learning Needs Hidden State (Or Much Faster Convergence)

Differential privacy analysis of randomized learning algorithms typicall...
research
10/03/2019

Recognizing the Tractability in Big Data Computing

Due to the limitation on computational power of existing computers, the ...
research
10/15/2014

Complexity Issues and Randomization Strategies in Frank-Wolfe Algorithms for Machine Learning

Frank-Wolfe algorithms for convex minimization have recently gained cons...

Please sign up or login with your details

Forgot password? Click here to reset