University of Cambridge > Talks.cam > Inference Group > Topic Modeling: Beyond Bag-of-Words

Topic Modeling: Beyond Bag-of-Words

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact Phil Cowans.

Some models of textual corpora employ text generation methods involving n-gram statistics, while others use latent topic variables inferred using the ``bag-of-words’’ assumption, in which word order is ignored. Previously, these methods have not been combined. In this talk, I present a hierarchical generative probabilistic model that incorporates both $n$-gram statistics and latent topic variables, by extending a unigram topic model to include properties of a hierarchical Dirichlet bigram language model.

This talk is part of the Inference Group series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2020 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity