Hierarchical Probabilistic Neural Network Language Model


Paper Link

Authors: Frederic Morin, Yoshua Bengio

Aim

Various neural network architecture have been used for Language modeling. With the emergence of word embeddings, these models have been successfully applied. However these models are quite slow in comparison to traditional n-gram models. To speed-up training and prediction, hierarchical the decomposition of the conditional probabilities that yields a speed-up by a factor about 200.

Introduction

In n-gram model as the size of vocabulary increases, the dictionary sizes with all combinations shoots up very sharply.

Challenges

  • Traversal across all combination even if a small fraction of update is done.
  • Similar object should have the same probability which is not possible if knowledge free representation is used.
  • Long training time and huge computation is needed.
The objective of this paper is thus to propose a much faster variant of the neural probabilistic language model.
Follow @yardstick17
By @Amit Kushwaha in
Tags :

Comments !