Training Neural Networks with Fixed Sparse Masks

11/18/2021
by   Yi-Lin Sung, et al.
0

During typical gradient-based training of deep neural networks, all of the model's parameters are updated at each iteration. Recent work has shown that it is possible to update only a small subset of the model's parameters during training, which can alleviate storage and communication requirements. In this paper, we show that it is possible to induce a fixed sparse mask on the model's parameters that selects a subset to update over many iterations. Our method constructs the mask out of the k parameters with the largest Fisher information as a simple approximation as to which parameters are most important for the task at hand. In experiments on parameter-efficient transfer learning and distributed training, we show that our approach matches or exceeds the performance of other methods for training with sparse updates while being more efficient in terms of memory usage and communication costs. We release our code publicly to promote further applications of our approach.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/18/2023

Calibrating the Rigged Lottery: Making All Tickets Reliable

Although sparse training has been successfully used in various resource-...
research
06/08/2023

A Gradient-based Approach for Online Robust Deep Neural Network Training with Noisy Labels

Learning with noisy labels is an important topic for scalable training i...
research
09/05/2019

Diversely Stale Parameters for Efficient Training of CNNs

The backpropagation algorithm is the most popular algorithm training neu...
research
11/18/2021

Merging Models with Fisher-Weighted Averaging

Transfer learning provides a way of leveraging knowledge from one task w...
research
11/17/2017

Training Simplification and Model Simplification for Deep Learning: A Minimal Effort Back Propagation Method

We propose a simple yet effective technique to simplify the training and...
research
03/29/2022

Kernel Modulation: A Parameter-Efficient Method for Training Convolutional Neural Networks

Deep Neural Networks, particularly Convolutional Neural Networks (ConvNe...
research
05/26/2021

Probabilistic Selective Encryption of Convolutional Neural Networks for Hierarchical Services

Model protection is vital when deploying Convolutional Neural Networks (...

Please sign up or login with your details

Forgot password? Click here to reset