On the State of the Art of Evaluation in Neural Language Models

07/18/2017
by   Gábor Melis, et al.
0

Ongoing innovations in recurrent neural network architectures have provided a steady influx of apparently state-of-the-art results on language modelling benchmarks. However, these have been evaluated using differing code bases and limited computational resources, which represent uncontrolled sources of experimental variation. We reevaluate several popular architectures and regularisation methods with large-scale automatic black-box hyperparameter tuning and arrive at the somewhat surprising conclusion that standard LSTM architectures, when properly regularised, outperform more recent models. We establish a new state of the art on the Penn Treebank and Wikitext-2 corpora, as well as strong baselines on the Hutter Prize dataset.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/18/2019

Language models and Automated Essay Scoring

In this paper, we present a new comparative study on automatic essay sco...
research
06/29/2021

Reliable and Fast Recurrent Neural Network Architecture Optimization

This article introduces Random Error Sampling-based Neuroevolution (RESN...
research
08/21/2023

Dataset Quantization

State-of-the-art deep neural networks are trained with large amounts (mi...
research
11/03/2022

Circling Back to Recurrent Models of Language

Just because some purely recurrent models suffer from being hard to opti...
research
03/10/2020

Large-Scale Evaluation of Keyphrase Extraction Models

Keyphrase extraction models are usually evaluated under different, not d...
research
06/11/2021

HPO-B: A Large-Scale Reproducible Benchmark for Black-Box HPO based on OpenML

Hyperparameter optimization (HPO) is a core problem for the machine lear...
research
11/09/2019

On Architectures for Including Visual Information in Neural Language Models for Image Description

A neural language model can be conditioned into generating descriptions ...

Please sign up or login with your details

Forgot password? Click here to reset