Investigation and Analysis of Hyper and Hypo neuron pruning to selectively update neurons during Unsupervised Adaptation

01/06/2020
by   Vikramjit Mitra, et al.
18

Unseen or out-of-domain data can seriously degrade the performance of a neural network model, indicating the model's failure to generalize to unseen data. Neural net pruning can not only help to reduce a model's size but can improve the model's generalization capacity as well. Pruning approaches look for low-salient neurons that are less contributive to a model's decision and hence can be removed from the model. This work investigates if pruning approaches are successful in detecting neurons that are either high-salient (mostly active or hyper) or low-salient (barely active or hypo), and whether removal of such neurons can help to improve the model's generalization capacity. Traditional blind adaptation techniques update either the whole or a subset of layers, but have never explored selectively updating individual neurons across one or more layers. Focusing on the fully connected layers of a convolutional neural network (CNN), this work shows that it may be possible to selectively adapt certain neurons (consisting of the hyper and the hypo neurons) first, followed by a full-network fine tuning. Using the task of automatic speech recognition, this work demonstrates how the removal of hyper and hypo neurons from a model can improve the model's performance on out-of-domain speech data and how selective neuron adaptation can ensure improved performance when compared to traditional blind model adaptation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/25/2020

Neuron Merging: Compensating for Pruned Neurons

Network pruning is widely used to lighten and accelerate neural network ...
research
11/16/2021

Neuron-based Pruning of Deep Neural Networks with Better Generalization using Kronecker Factored Curvature Approximation

Existing methods of pruning deep neural networks focus on removing unnec...
research
07/12/2016

Network Trimming: A Data-Driven Neuron Pruning Approach towards Efficient Deep Architectures

State-of-the-art neural networks are getting deeper and wider. While the...
research
04/18/2018

Understanding Individual Neuron Importance Using Information Theory

In this work, we characterize the outputs of individual neurons in a tra...
research
12/02/2019

TX-Ray: Quantifying and Explaining Model-Knowledge Transfer in (Un-)Supervised NLP

While state-of-the-art NLP explainability (XAI) methods focus on supervi...
research
06/24/2010

Active Sites model for the B-Matrix Approach

This paper continues on the work of the B-Matrix approach in hebbian lea...
research
08/31/2017

Leveraging Deep Neural Network Activation Entropy to cope with Unseen Data in Speech Recognition

Unseen data conditions can inflict serious performance degradation on sy...

Please sign up or login with your details

Forgot password? Click here to reset