Scaling Recurrent Neural Network Language Models

02/02/2015
by   Will Williams, et al.
0

This paper investigates the scaling properties of Recurrent Neural Network Language Models (RNNLMs). We discuss how to train very large RNNs on GPUs and address the questions of how RNNLMs scale with respect to model size, training-set size, computational costs and memory. Our analysis shows that despite being more costly to train, RNNLMs obtain much lower perplexities on standard benchmarks than n-gram models. We train the largest known RNNs and present relative word error rates gains of 18 the new lowest perplexities on the recently released billion word language modelling benchmark, 1 BLEU point gain on machine translation and a 17 relative hit rate gain in word prediction.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/21/2015

BlackOut: Speeding up Recurrent Neural Network Language Models With Very Large Vocabularies

We propose BlackOut, an approximation algorithm to efficiently train mas...
research
06/22/2019

Evaluating Computational Language Models with Scaling Properties of Natural Language

In this article, we evaluate computational models of natural language wi...
research
11/27/2017

Slim Embedding Layers for Recurrent Neural Language Models

Recurrent neural language models are the state-of-the-art models for lan...
research
01/03/2016

Contrastive Entropy: A new evaluation metric for unnormalized language models

Perplexity (per word) is the most widely used metric for evaluating lang...
research
06/22/2018

Evaluating language models of tonal harmony

This study borrows and extends probabilistic language models from natura...
research
09/26/2021

Curb Your Carbon Emissions: Benchmarking Carbon Emissions in Machine Translation

In recent times, there has been definitive progress in the field of NLP,...

Please sign up or login with your details

Forgot password? Click here to reset