# A Simple Language Model based on PMI Matrix Approximations

###### Abstract

In this study, we introduce a new approach for learning language models by training them to estimate word-context pointwise mutual information (PMI), and then deriving the desired conditional probabilities from PMI at test time. Specifically, we show that with minor modifications to word2vec’s algorithm, we get principled language models that are closely related to the well-established Noise Contrastive Estimation (NCE) based language models. A compelling aspect of our approach is that our models are trained with the same simple negative sampling objective function that is commonly used in word2vec to learn word embeddings.

A Simple Language Model based on PMI Matrix Approximations

Oren Melamud IBM Research Yorktown Heights, NY, USA oren.melamud@ibm.com Ido Dagan Computer Science Dept. Bar-Ilan University, Israel dagan@cs.biu.ac.il Jacob Goldberger Faculty of Engineering Bar-Ilan University, Israel jacob.goldberger@biu.ac.il

## 1 Introduction

Language models (LMs) learn to estimate the probability of a word given a context of preceding words. Recurrent Neural Network (RNN) language models recently outperformed traditional -gram LMs across a range of tasks (Jozefowicz et al., 2016). However, an important practical issue associated with such neural-network LMs is the high computational cost incurred. The key factor that limits the scalability of traditional neural LMs is the computation of the normalization term in the softmax output layer, whose cost is linearly proportional to the size of the word vocabulary.

Several methods have been proposed to cope with this scaling issue by replacing the softmax with a more computationally efficient component at train time.^{1}^{1}1An alternative recent approach for coping with large word vocabularies is to represent words as compositions of sub-word units, such as individual characters. This approach has notable merits (Jozefowicz et al., 2016; Sennrich et al., 2016), but is out of the scope of this paper.
These include importance sampling (Bengio and et al, 2003), hierarchical softmax (Minh and Hinton, 2008), BlackOut (Ji et al., 2016) and Noise Contrastive Estimation (NCE) (Gutmann and Hyvarinen, 2012).
NCE has been applied to train neural LMs with large vocabularies (Mnih and Teh, 2012) and more recently was also successfully used to train LSTM-RNN LMs (Vaswani et al., 2013; Chen et al., 2015; Zoph et al., 2016). NCE-based language models achieved near state-of-the-art performance on language modeling tasks (Jozefowicz et al., 2016; Chen et al., 2016), and as we later show, are closely related to the method presented in this paper.

Continuous word embeddings were initially introduced as a ‘by-product’ of learning neural language models (Bengio and et al, 2003). However, they were later adopted in many other NLP tasks, and the most popular recent word embedding learning models are no longer proper language models. In particular, the skip-gram with negative sampling (NEG) embedding algorithm (Mikolov et al., 2013) as implemented in the word2vec toolkit, has become one of the most popular such models today. This is largely attributed to its scalability to huge volumes of data, which is critical for learning high-quality embeddings. Recently, Levy and Goldberg (2014) offered a motivation for the NEG objective function, showing that by maximizing this function, the skip-gram algorithm implicitly attempts to factorize a word-context pointwise mutual information (PMI) matrix. Melamud and Goldberger (2017) rederived this result by offering an information-theory interpretation of NEG.

The NEG objective function is considered a simplification of the NCE’s objective, unsuitable for learning language models (Dyer, 2014). However, in this study, we show that despite its simplicity, it can be used in a principled way to effectively train a language model, based on PMI matrix factorization. More specifically, we use NEG to train a model for estimating the PMI between words and their preceding contexts, and then derive conditional probabilities from PMI at test time. The obtained PMI-LM can be viewed as a simple variant of word2vec’s algorithm, where the context of a predicted word is the preceding sequence of words, rather than a single word within a context window (skip-gram), or a bag-of-context-words (CBOW).

Our analysis shows that the proposed PMI-LM is very closely related to NCE language models (NCE-LMs). Similar to NCE-LMs, PMI-LM avoids the dependency of train run-time on the size of the word vocabulary by sampling from a negative (noise) distribution. Furthermore, conveniently, it also has a notably more simplified objective function formulation inherited from word2vec, which allows it to avoid the heuristic components and initialization procedures used in various implementations of NCE language models (Vaswani et al., 2013; Chen et al., 2015; Zoph et al., 2016).

Finally, we report on a perplexity evaluation of PMI and NCE language models on two standard language modeling datasets. The evaluation yielded comparable results, supporting our theoretical analysis.

## 2 NCE-based Language Modeling

Noise Contrastive Estimation (NCE) has recently been used to learn language models efficiently. NCE transforms the parameter learning problem into a binary classifier training problem. Let be the probability of a word given a context that represents its entire preceding context, and let be a ‘noise’ word distribution (e.g. a unigram distribution). The NCE approach assumes that the word is sampled from a mixture distribution such that the noise samples are times more frequent than samples from the ‘true’ distribution . Let be a binary random variable such that and correspond to a noise sample and a true sample, respectively, i.e. and . Assume the distribution has the following parametric form:

(1) |

such that and are vector representations of the word and its context . Applying Bayes rule, it can be easily verified that:

(2) |

where is the sigmoid function.

NCE uses Eq. (2) and the following objective function to train a binary classifier that decides which distribution was used to sample :

(3) |

such that go over all the word-context co-occurrences in the learning corpus and are ‘noise’ samples drawn from the word unigram distribution.

Note that the normalization factor is not a free parameter and to obtain its value, one needs to compute for each context , where is the word vocabulary. This computation is typically not feasible due to the large vocabulary size and the exponentially large number of possible contexts and therefore it was heuristically circumvented by prior work. Mnih and Teh (2012) found empirically that setting didn’t hurt the performance (see also discussion in (Andreas and Klein, 2015)). Chen et al. (2015) reported that setting gave them the best results. Recent works (Vaswani et al., 2013; Zoph et al., 2016) used and also initialized NCE’s bias term from Eq. (2) to . They reported that without these heuristics the training procedure did not converge to a meaningful model.

In the following section, we describe our proposed language model, which is derived from word2vec’s interpretation as a low-rank PMI matrix approximation. Interestingly, this model turns out to be a close variant of NCE language models, but with a simplified objective function that avoids the need for the normalization factor and the bias terms.

## 3 PMI-based Language Modeling

The skip-gram negative sampling word embedding algorithm represents each word and each context word as -dimensional vectors, with the purpose that words that are “similar” to each other will have similar vector representations. The algorithm optimizes the following NEG objective function (Mikolov et al., 2013):

(4) |

such that go over all the word-context co-occurrences in the learning corpus , are words independently sampled from the word unigram distribution, is the embedding of and is the sigmoid function. The objective function can be viewed as a log-likelihood function of a binary logistic regression classifier that treats a sample from a joint word-context distribution as a positive instance, and two independent samples from the word and context unigram distributions as a negative instance, while is the proportion between negative and positive instances. Levy and Goldberg (2014) showed that this objective function achieves its maximal value when for every word-context pair :

(5) |

where is the word-context PMI matrix. Actually achieving this maximal value is typically infeasible, since the embedding dimensionality is intentionally limited. Therefore, learning word and context embeddings that optimize skip-gram’s NEG objective function (4) can be viewed as finding a low-rank approximation of the word-context PMI matrix. An explicit expression of the approximation criterion optimized by the skip-gram algorithm can be found in (Melamud and Goldberger, 2017).

Our study is based on two simple observations regarding this finding of Levy and Goldberg (2014). First, Equation (5) can be reformulated as follows to derive an estimate of the conditional distribution :

(6) |

where the constant is dropped since is a distribution. Second, while the above analysis had been originally applied to the case of word-context joint distributions , it is easy to see that the PMI matrix approximation analysis also holds for every Euclidean embedding of a joint distribution of any two given random variables and . In particular, we note that it holds for word-context joint distributions , where is a single word, but represents its entire preceding context, rather than just a single context word, and is a vector representation of this entire context. Altogether, this allows us to use word2vec’s NEG objective function (4) to approximate the language modeling conditional probability (6), with being the entire preceding context of the predicted word .

Training objective function | Test probability estimate | |
---|---|---|

NCE-LM | ||

PMI-LM |

We next describe the design details of the proposed PMI-based language modeling. We use a simple lookup table for the word representation , and an LSTM recurrent neural network to obtain a low dimensional representation of the entire preceding context . These representations are trained to maximize the NEG objective in Eq. (4), where this time goes over every word token in the corpus, and is its preceding context. We showed above that optimizing this objective seeks to obtain the best low-dimensional approximation of the PMI matrix associated with the joint distribution of the word and its preceding context (Eq. (5)). Hence, based on Eq. (6), for a reasonable embedding dimensionality and a good model for representing the preceding context, we expect to be a good estimate of the language modeling conditional distribution.

At test time, to obtain a proper distribution, we perform a normalization operation as done by all other comparable models. The train and test steps of the proposed language modeling algorithm are shown in algorithm box 1.

Note that while the NCE approach (1) learns to explicitly estimate normalized conditional distributions, our model learns to approximate the PMI matrix. Hence, we have no real motivation to include additional learned normalization parameters, as considered in comparable NCE language models (Mnih and Teh, 2012; Zoph et al., 2016).

The NEG and NCE objective functions share a similar form:

(7) |

with the differences summarized in Table 1. The comparison shows that PMI-LM’s NEG objective function is much simpler. Furthermore, due to the component in NCE’s objective function, its input to the sigmoid function is sensitive to the variable values in the unigram distribution, and therefore potentially more difficult to concentrate around zero with low variance to facilitate effective back-propagation. This may explain heuristics used by prior work for initializing the values of (Vaswani et al., 2013; Zoph et al., 2016).

## 4 Experiments

The goal of the evaluation described in this section is to empirically establish PMI-LM as a sound language model. We do so by comparing its performance with the well-established NCE-LM, using the popular perplexity measure on two standard datasets, under the same terms. We describe our hyperparameter choices below and stress that for a fair comparison, we followed prior best practices and avoided hyperparameter optimization in favor of PMI-LM. All of the models described hereafter were implemented using the Chainer toolkit (Tokui et al., 2015).

For our NCE baseline, we used the heuristics that worked well in (Vaswani et al., 2013; Zoph et al., 2016), initializing NCE’s bias term from Eq. (2) to , where is the word vocabulary, and using .

The first dataset we used is a version of the Penn Tree Bank (PTB), commonly used to evaluate language models.^{2}^{2}2Available from Tomas Mikolov at: http://www.fit.vutbr.cz/~imikolov/rnnlm/simple-examples.tgz It consists of 929K training words, 73K validation words and 82K test words with a 10K
word vocabulary. To build and train the compared models in this setting, we followed the work of Zaremba et al. Zaremba et al. (2014), who achieved excellent results on this dataset. Specifically, we used a 2-layer 300-hidden-units LSTM with a 50% dropout ratio to represent the preceding (left-side) context of a predicted word.^{3}^{3}3Zaremba et al. Zaremba et al. (2014) used larger models with more units and also applied dropout to the output of the top LSTM layer, which we did not. We represented end-of-sentence as a special eos token and predicted this token like any other word. During training, we performed truncated back-propagation-through-time, unrolling the LSTM for 20 steps at a time without ever resetting the LSTM state. We trained our model for 39 epochs using Stochastic Gradient Descent (SGD) with a learning rate of 1, which is decreased by a factor of 1.2 after every epoch starting after epoch 6. We clipped the norms of the gradient to 5 and used a mini-batch size of 20. We set the negative sampling parameter to following Zoph et al. Zoph et al. (2016), who showed highly competitive performance with NCE LMs trained with this number of samples.

As the second dataset, we used the much larger WMT 1B-word benchmark introduced by Chelba et al. Chelba et al. (2013). This dataset comprises about 0.8B training words and has a held-out set partitioned into 50 subsets. The test set is the first subset in the held-out, comprising 159K words, including the eos tokens. We used the second subset as the validation set with 165K words. The original vocabulary size of this dataset is 0.8M words after converting all words that occur less than 3 times in the corpus to an unk token. However, we followed previous works (Williams et al., 2015; Ji et al., 2016) and trimmed the vocabulary further down to the top 64K most frequent words in order to successfully fit a neural model to this data using reasonably modest compute resources. To build and train our models, we used a similar method to the one used with PTB, with the following differences. We used a single-layer 512-hidden-unit LSTM to represent the preceding context. We followed Jozefowicz et al. Jozefowicz et al. (2016), who found a 10% dropout rate to be sufficient for relatively small models fitted to this large training corpus. We trained our model for only one epoch using the Adam optimizer (Kingma and Ba, 2014) with default parameters, which we found to converge more quickly and effectively than SGD. We used a mini-batch size of 1000.

PMI-LM | NCE-LM | |
---|---|---|

PTB | 98.35 | 104.33 |

WMT | 65.84 | 69.28 |

The perplexity results achieved by the compared models appear in Table 2. As can be seen, the performance of our PMI-LM is competitive, slightly outperforming the NCE-LM on both test sets. To put these numbers in a broader context, we note that state-of-the-art results on these datasets are notably better. For example, on the small PTB test set, Zaremba et al. (2014) achieved 78.4 perplexity with a larger LSTM model and using the more costly softmax component. On the larger WMT dataset, Jozefowicz et al. (2016) achieved 46.1 and 43.7 perplexity numbers using NCE and importance sampling respectively, and with much larger LSTM models trained over the full vocabulary, rather than our trimmed one. They also achieved 23.7 with an ensemble method, which is the best result on this dataset to date. Yet, as intended, we argue that our experimental results affirm the claim that PMI-LM is a sound language model on par with NCE-LM.

## 5 Conclusions

In this work, we have shown that word2vec’s negative sampling objective function, popularized in the context of learning word representations, can also be used to effectively learn parametric language models. These language models are closely related to NCE language models, but utilize a simpler, potentially more robust objective function. More generally, our theoretical analysis shows that any word2vec model trained with negative sampling can be used in a principled way to estimate the conditional distribution , by following our proposed procedure at test time.

## Acknowledgments

This work is supported by the Intel Collaborative Research Institute for Computational Intelligence (ICRI-CI).

## References

- Andreas and Klein (2015) J. Andreas and D. Klein. 2015. When and why are log-linear models self-normalizing? In NAACL.
- Bengio and et al (2003) Y. Bengio and J. Senecal et al. 2003. Quick training of probabilistic neural nets by importance sampling. In AISTATS.
- Chelba et al. (2013) C. Chelba, T. Mikolov, M. Schuster, Q. Ge, T. Brants, P. Koehn, and T. Robinson. 2013. One billion word benchmark for measuring progress in statistical language modeling. arXiv preprint arXiv:1312.3005.
- Chen et al. (2016) W. Chen, D. Grangier, and M. Auli. 2016. Strategies for training large vocabulary neural language models. CoRR, abs/1512.04906.
- Chen et al. (2015) X. Chen, X. Liu, M. Gales, and P. C. Woodland. 2015. Recurrent neural network language model training with noise contrastive estimation for speech recognition. In ICASSP.
- Dyer (2014) C. Dyer. 2014. Notes on noise contrastive estimation and negative sampling. arXiv preprint arXiv:1410.8251.
- Gutmann and Hyvarinen (2012) M. U. Gutmann and A. Hyvarinen. 2012. Noise-contrastive estimation of unnormalized statistical models, with applications to natural image statistics. Journal of Machine Learning Research, 13:307–361.
- Ji et al. (2016) S. Ji, S. Vishwanathan, N. Satish, A. Nadathur, J. Michael, and P. Dubey. 2016. Blackout: Speeding up recurrent neural network language models with very large vocabularies. ICLR.
- Jozefowicz et al. (2016) R. Jozefowicz, O. Vinyals, M. Schuster, N. Shazeer, and Y. Wu. 2016. Exploring the limits of language modeling. arXiv preprint arXiv:1602.02410.
- Kingma and Ba (2014) D. Kingma and J. Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980.
- Levy and Goldberg (2014) O. Levy and Y. Goldberg. 2014. Neural word embedding as implicit matrix factorization. In Advances in Neural Information Processing Systems.
- Melamud and Goldberger (2017) O. Melamud and J. Goldberger. 2017. Information-theory interpretation of the skip-gram negative-sampling objective function. In Proceedings of ACL.
- Mikolov et al. (2013) T. Mikolov, I. Sutskever, K. Chen, G. Corrado, and J. Dean. 2013. Distributed representations of words and phrases and their compositionality. In Advances in Neural Information Processing Systems.
- Minh and Hinton (2008) A. Minh and G. E. Hinton. 2008. A scalable hierarchical distributed language model. In Advances in Neural Information Processing Systems.
- Mnih and Teh (2012) A. Mnih and Y. W. Teh. 2012. A fast and simple algorithm for training neural probabilistic language models. In ICML.
- Sennrich et al. (2016) R. Sennrich, B. Haddow, and A. Birch. 2016. Neural machine translation of rare words with subword units. CoRR, abs/1508.07909.
- Tokui et al. (2015) S. Tokui, K. Oono, S. Hido, and J. Clayton. 2015. Chainer: a next-generation open source framework for deep learning. In Workshop on Machine Learning Systems (LearningSys) in The 29th Annual Conference on Neural Information Processing Systems.
- Vaswani et al. (2013) A. Vaswani, Y. Zhao, V. Fossum, and D. Chiang. 2013. Decoding with large-scale neural language models improves translation. In EMNLP.
- Williams et al. (2015) W. Williams, N. Prasad, D. Mrva, T. Ash, and T. Robinson. 2015. Scaling recurrent neural network language models. ICASSP.
- Zaremba et al. (2014) W. Zaremba, I. Sutskever, and O. Vinyals. 2014. Recurrent neural network regularization. arXiv preprint arXiv:1409.2329.
- Zoph et al. (2016) B. Zoph, A. Vaswani, J. May, and K. Knight. 2016. Simple, fast noise-contrastive estimation for large RNN vocabularies. In NAACL.