Topic Modeling: Beyond Bag-of-Words
Add to your list(s)
Download to your calendar using vCal
If you have a question about this talk, please contact Phil Cowans.
Some models of textual corpora employ text generation methods involving n-gram statistics, while others use latent topic variables inferred using the ``bag-of-words’’ assumption, in which word order is ignored. Previously, these methods have not been combined. In this talk, I present a hierarchical generative probabilistic model that incorporates both $n$-gram statistics and latent topic variables, by extending a unigram topic model to include properties of a hierarchical Dirichlet bigram language model.
This talk is part of the Inference Group series.
This talk is included in these lists:
Note that ex-directory lists are not shown.
|