Normalization Before Shaking Toward Learning Symmetrically Distributed Representation Without Margin in Speech Emotion Recognition

08/02/2018
by   Che-Wei Huang, et al.
0

Regularization is crucial to the success of many practical deep learning models, in particular in a more often than not scenario where there are only a few to a moderate number of accessible training samples. In addition to weight decay, data augmentation and dropout, regularization based on multi-branch architectures, such as Shake-Shake regularization, has been proven successful in many applications and attracted more and more attention. However, beyond model-based representation augmentation, it is unclear how Shake-Shake regularization helps to provide further improvement on classification tasks, let alone the baffling interaction between batch normalization and shaking. In this work, we present our investigation on Shake-Shake regularization, drawing connections to the vicinal risk minimization principle and discriminative feature learning in face verification. Furthermore, we identify a strong resemblance between batch normalized residual blocks and batch normalized recurrent neural networks, where both of them share a similar convergence behavior, which could be mitigated by a proper initialization of batch normalization. Based on the findings, our experiments on speech emotion recognition demonstrate simultaneously an improvement on the classification accuracy and a reduction on the generalization gap both with statistical significance.

READ FULL TEXT
research
11/09/2022

A Comparative Study of Data Augmentation Techniques for Deep Learning Based Emotion Recognition

Automated emotion recognition in speech is a long-standing problem. Whil...
research
04/18/2018

Shaking Acoustic Spectral Sub-bands Can Better Regularize Learning in Affective Computing

In this work, we investigate a recently proposed regularization techniqu...
research
07/19/2019

Post-synaptic potential regularization has potential

Improving generalization is one of the main challenges for training deep...
research
01/05/2020

Self-Orthogonality Module: A Network Architecture Plug-in for Learning Orthogonal Filters

In this paper, we investigate the empirical impact of orthogonality regu...
research
07/08/2016

Adjusting for Dropout Variance in Batch Normalization and Weight Initialization

We show how to adjust for the variance introduced by dropout with correc...
research
05/08/2023

LABO: Towards Learning Optimal Label Regularization via Bi-level Optimization

Regularization techniques are crucial to improving the generalization pe...
research
03/20/2019

Deep Octonion Networks

Deep learning is a research hot topic in the field of machine learning. ...

Please sign up or login with your details

Forgot password? Click here to reset