Characterizing the hyper-parameter space of LSTM language models for mixed context applications

12/08/2017
by   Victor Akinwande, et al.
0

Applying state of the art deep learning models to novel real world datasets gives a practical evaluation of the generalizability of these models. Of importance in this process is how sensitive the hyper parameters of such models are to novel datasets as this would affect the reproducibility of a model. We present work to characterize the hyper parameter space of an LSTM for language modeling on a code-mixed corpus. We observe that the evaluated model shows minimal sensitivity to our novel dataset bar a few hyper parameters.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/23/2021

LSTM Hyper-Parameter Selection for Malware Detection: Interaction Effects and Hierarchical Selection Approach

Long-Short-Term-Memory (LSTM) networks have shown great promise in artif...
research
12/19/2019

CNN-LSTM models for Multi-Speaker Source Separation using Bayesian Hyper Parameter Optimization

In recent years there have been many deep learning approaches towards th...
research
03/27/2019

Echo State Networks with Self-Normalizing Activations on the Hyper-Sphere

Among the various architectures of Recurrent Neural Networks, Echo State...
research
12/26/2020

Assessment of the Relative Importance of different hyper-parameters of LSTM for an IDS

Recurrent deep learning language models like the LSTM are often used to ...
research
11/06/2021

What augmentations are sensitive to hyper-parameters and why?

We apply augmentations to our dataset to enhance the quality of our pred...
research
05/10/2014

A Hybrid Monte Carlo Architecture for Parameter Optimization

Much recent research has been conducted in the area of Bayesian learning...
research
11/24/2019

Stage-based Hyper-parameter Optimization for Deep Learning

As deep learning techniques advance more than ever, hyper-parameter opti...

Please sign up or login with your details

Forgot password? Click here to reset