Neural Rejuvenation: Improving Deep Network Training by Enhancing Computational Resource Utilization

12/02/2018
by   Siyuan Qiao, et al.
0

In this paper, we study the problem of improving computational resource utilization of neural networks. Deep neural networks are usually over-parameterized for their tasks in order to achieve good performances, thus are likely to have underutilized computational resources. This observation motivates a lot of research topics, e.g. network pruning, architecture search, etc. As models with higher computational costs (e.g. more parameters or more computations) usually have better performances, we study the problem of improving the resource utilization of neural networks so that their potentials can be further realized. To this end, we propose a novel optimization method named Neural Rejuvenation. As its name suggests, our method detects dead neurons and computes resource utilization in real time, rejuvenates dead neurons by resource reallocation and reinitialization, and trains them with new training schemes. By simply replacing standard optimizers with Neural Rejuvenation, we are able to improve the performances of neural networks by a very large margin while using similar training efforts and maintaining their original resource usages.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/13/2021

Self-Reorganizing and Rejuvenating CNNs for Increasing Model Capacity Utilization

In this paper, we propose self-reorganizing and rejuvenating convolution...
research
08/18/2023

Adaptive Timers and Buffer Optimization for Layer-2 Protocols in 5G Non-Terrestrial Networks

Interest in the integration of Terrestrial Networks (TN) and Non-Terrest...
research
11/18/2016

NoiseOut: A Simple Way to Prune Neural Networks

Neural networks are usually over-parameterized with significant redundan...
research
05/24/2022

DNNAbacus: Toward Accurate Computational Cost Prediction for Deep Neural Networks

Deep learning is attracting interest across a variety of domains, includ...
research
04/25/2021

Balancing Accuracy and Latency in Multipath Neural Networks

The growing capacity of neural networks has strongly contributed to thei...
research
07/30/2023

An Efficient Approach to Mitigate Numerical Instability in Backpropagation for 16-bit Neural Network Training

In this research, we delve into the intricacies of the numerical instabi...
research
09/27/2021

Consistency Training of Multi-exit Architectures for Sensor Data

Deep neural networks have become larger over the years with increasing d...

Please sign up or login with your details

Forgot password? Click here to reset