Finding trainable sparse networks through Neural Tangent Transfer

06/15/2020
by   Tianlin Liu, et al.
0

Deep neural networks have dramatically transformed machine learning, but their memory and energy demands are substantial. The requirements of real biological neural networks are rather modest in comparison, and one feature that might underlie this austerity is their sparse connectivity. In deep learning, trainable sparse networks that perform well on a specific task are usually constructed using label-dependent pruning criteria. In this article, we introduce Neural Tangent Transfer, a method that instead finds trainable sparse networks in a label-free manner. Specifically, we find sparse networks whose training dynamics, as characterized by the neural tangent kernel, mimic those of dense networks in function space. Finally, we evaluate our label-agnostic approach on several standard classification tasks and show that the resulting sparse networks achieve higher classification performance while converging faster.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/25/2020

Data Parallelism in Training Sparse Neural Networks

Network pruning is an effective methodology to compress large neural net...
research
03/28/2016

Sparse Activity and Sparse Connectivity in Supervised Learning

Sparseness is a useful regularizer for learning in a wide range of appli...
research
05/12/2020

Modularizing Deep Learning via Pairwise Learning With Kernels

By redefining the conventional notions of layers, we present an alternat...
research
06/17/2021

On the training of sparse and dense deep neural networks: less parameters, same performance

Deep neural networks can be trained in reciprocal space, by acting on th...
research
07/07/2021

Self-organized criticality in neural networks

We demonstrate, both analytically and numerically, that learning dynamic...
research
03/08/2022

Dual Lottery Ticket Hypothesis

Fully exploiting the learning capacity of neural networks requires overp...
research
10/28/2017

Trainable back-propagated functional transfer matrices

Connections between nodes of fully connected neural networks are usually...

Please sign up or login with your details

Forgot password? Click here to reset