A Framework for Parallel and Distributed Training of Neural Networks

10/24/2016
by   Simone Scardapane, et al.
0

The aim of this paper is to develop a general framework for training neural networks (NNs) in a distributed environment, where training data is partitioned over a set of agents that communicate with each other through a sparse, possibly time-varying, connectivity pattern. In such distributed scenario, the training problem can be formulated as the (regularized) optimization of a non-convex social cost function, given by the sum of local (non-convex) costs, where each agent contributes with a single error term defined with respect to its local dataset. To devise a flexible and efficient solution, we customize a recently proposed framework for non-convex optimization over networks, which hinges on a (primal) convexification-decomposition technique to handle non-convexity, and a dynamic consensus procedure to diffuse information among the agents. Several typical choices for the training criterion (e.g., squared loss, cross entropy, etc.) and regularization (e.g., ℓ_2 norm, sparsity inducing penalties, etc.) are included in the framework and explored along the paper. Convergence to a stationary solution of the social non-convex problem is guaranteed under mild assumptions. Additionally, we show a principled way allowing each agent to exploit a possible multi-core architecture (e.g., a local cloud) in order to parallelize its local optimization step, resulting in strategies that are both distributed (across the agents) and parallel (inside each agent) in nature. A comprehensive set of experimental results validate the proposed approach.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/24/2017

A randomized primal distributed algorithm for partitioned and big-data non-convex optimization

In this paper we consider a distributed optimization scenario in which t...
research
06/15/2017

Stochastic Training of Neural Networks via Successive Convex Approximations

This paper proposes a new family of algorithms for training neural netwo...
research
06/28/2015

Beating the Perils of Non-Convexity: Guaranteed Training of Neural Networks using Tensor Methods

Training neural networks is a challenging non-convex optimization proble...
research
04/30/2020

Distributed Stochastic Nonconvex Optimization and Learning based on Successive Convex Approximation

We study distributed stochastic nonconvex optimization in multi-agent ne...
research
05/31/2019

Deep Learning for Distributed Optimization: Applications to Wireless Resource Management

This paper studies a deep learning (DL) framework to solve distributed n...
research
03/18/2019

Annealing for Distributed Global Optimization

The paper proves convergence to global optima for a class of distributed...
research
08/14/2023

Self-Healing First-Order Distributed Optimization with Packet Loss

We describe SH-SVL, a parameterized family of first-order distributed op...

Please sign up or login with your details

Forgot password? Click here to reset