GIANT: Globally Improved Approximate Newton Method for Distributed Optimization

09/11/2017
by   Shusen Wang, et al.
0

For distributed computing environments, we consider the canonical machine learning problem of empirical risk minimization (ERM) with quadratic regularization, and we propose a distributed and communication-efficient Newton-type optimization method. At every iteration, each worker locally finds an Approximate NewTon (ANT) direction, and then it sends this direction to the main driver. The driver, then, averages all the ANT directions received from workers to form a Globally Improved ANT (GIANT) direction. GIANT naturally exploits the trade-offs between local computations and global communications in that more local computations result in fewer overall rounds of communications. GIANT is highly communication efficient in that, for d-dimensional data uniformly distributed across m workers, it has 4 or 6 rounds of communication and O (d m) communication complexity per iteration. Theoretically, we show that GIANT's convergence rate is faster than first-order methods and existing distributed Newton-type methods. From a practical point-of-view, a highly beneficial feature of GIANT is that it has only one tuning parameter---the iterations of the local solver for computing an ANT direction. This is indeed in sharp contrast with many existing distributed Newton-type methods, as well as popular first-order methods, which have several tuning parameters, and whose performance can be greatly affected by the specific choices of such parameters. In this light, we empirically demonstrate the superior performance of GIANT compared with other competing methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/10/2020

DONE: Distributed Newton-type Method for Federated Edge Learning

There is growing interest in applying distributed machine learning to ed...
research
01/01/2015

Communication-Efficient Distributed Optimization of Self-Concordant Empirical Loss

We consider distributed convex optimization problems originated from sam...
research
08/17/2023

Dual Gauss-Newton Directions for Deep Learning

Inspired by Gauss-Newton-like methods, we study the benefit of leveragin...
research
12/30/2013

Communication Efficient Distributed Optimization using an Approximate Newton-type Method

We present a novel Newton-type method for distributed optimization, whic...
research
08/20/2021

L-DQN: An Asynchronous Limited-Memory Distributed Quasi-Newton Method

This work proposes a distributed algorithm for solving empirical risk mi...
research
07/13/2022

Communication-efficient Distributed Newton-like Optimization with Gradients and M-estimators

In modern data science, it is common that large-scale data are stored an...
research
07/22/2019

Practical Newton-Type Distributed Learning using Gradient Based Approximations

We study distributed algorithms for expected loss minimization where the...

Please sign up or login with your details

Forgot password? Click here to reset