Skip-gram Language Modeling Using Sparse Non-negative Matrix Probability Estimation

12/03/2014
by   Noam Shazeer, et al.
0

We present a novel family of language model (LM) estimation techniques named Sparse Non-negative Matrix (SNM) estimation. A first set of experiments empirically evaluating it on the One Billion Word Benchmark shows that SNM n-gram LMs perform almost as well as the well-established Kneser-Ney (KN) models. When using skip-gram features the models are able to match the state-of-the-art recurrent neural network (RNN) LMs; combining the two modeling techniques yields the best known result on the benchmark. The computational advantages of SNM over both maximum entropy and RNN LM estimation are probably its main strength, promising an approach that has the same flexibility in combining arbitrary features effectively and yet should scale to very large amounts of data as gracefully as n-gram LMs do.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/05/2015

Multinomial Loss on Held-out Data for the Sparse Non-negative Matrix Language Model

We describe Sparse Non-negative Matrix (SNM) language model estimation u...
research
03/31/2017

N-gram Language Modeling using Recurrent Neural Network Estimation

We investigate the effective memory depth of RNN models by using them fo...
research
07/15/2019

Investigation on N-gram Approximated RNNLMs for Recognition of Morphologically Rich Speech

Recognition of Hungarian conversational telephone speech is challenging ...
research
06/13/2019

Character n-gram Embeddings to Improve RNN Language Models

This paper proposes a novel Recurrent Neural Network (RNN) language mode...
research
04/01/2018

Revisiting Skip-Gram Negative Sampling Model with Regularization

We revisit skip-gram negative sampling (SGNS), a popular neural-network ...
research
06/27/2020

Beneath (or beyond) the surface: Discovering voice-leading patterns with skip-grams

Recurrent voice-leading patterns like the Mi-Re-Do compound cadence (MRD...
research
03/26/2018

Connectionist Recommendation in the Wild

The aggregate behaviors of users can collectively encode deep semantic i...

Please sign up or login with your details

Forgot password? Click here to reset