Pushing the boundaries of parallel Deep Learning -- A practical approach

06/25/2018
by   Paolo Viviani, et al.
0

This work aims to assess the state of the art of data parallel deep neural network training, trying to identify potential research tracks to be exploited for performance improvement. Beside, it presents a design for a practical C++ library dedicated at implementing and unifying the current state of the art methodologies for parallel training in a performance-conscious framework, allowing the user to explore novel strategies without departing significantly from its usual work-flow.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/30/2016

A Theoretical Study of The Relationship Between Whole An ELM Network and Its Subnetworks

A biological neural network is constituted by numerous subnetworks and m...
research
09/22/2016

Distributed Training of Deep Neural Networks: Theoretical and Practical Limits of Parallel Scalability

This paper presents a theoretical analysis and practical evaluation of t...
research
03/22/2021

Groovy Parallel Patterns: A Process oriented Parallelization Library

A novel parallel patterns library, Groovy Parallel Patterns, is presente...
research
05/21/2018

Parameter Hub: a Rack-Scale Parameter Server for Distributed Deep Neural Network Training

Distributed deep neural network (DDNN) training constitutes an increasin...
research
04/12/2021

GPflux: A Library for Deep Gaussian Processes

We introduce GPflux, a Python library for Bayesian deep learning with a ...
research
12/19/2019

Multilevel Initialization for Layer-Parallel Deep Neural Network Training

This paper investigates multilevel initialization strategies for trainin...
research
09/17/2021

The Optimization of the Constant Flow Parallel Micropump Using RBF Neural Network

The objective of this work is to optimize the performance of a constant ...

Please sign up or login with your details

Forgot password? Click here to reset