A Hierarchical Bayesian Language Model based on Pitman-Yor Processes
Add to your list(s)
Download to your calendar using vCal
If you have a question about this talk, please contact Hanna Wallach.
Paper
(also Tech. report)
We propose a new hierarchical Bayesian n-gram model of natural languages. Our model makes use of a generalization of the commonly used Dirichlet distributions called Pitman-Yor processes which produce power-law distributions more closely resembling those in natural languages. We show that an approximation to the hierarchical Pitman-Yor language model recovers the exact formulation of interpolated Kneser-Ney, one of the best smoothing methods for n-gram language models. Experiments verify that our model gives cross entropy results superior to interpolated Kneser-Ney and comparable to modified Kneser-Ney.
This talk is part of the Machine Learning Journal Club series.
This talk is included in these lists:
Note that ex-directory lists are not shown.
|