Probabilistic Topic Modeling in Multilingual Settings: A Short Overview of Its Methodology and Applications
published: Jan. 11, 2013, recorded: December 2012, views: 5334
Report a problem or upload filesIf you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.
Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
Probabilistic topic models are unsupervised generative models that model document content as a two-step generation process, i.e., documents are observed as mixtures of latent topics, while topics are probability distributions over vocabulary words. Recently, a significant research effort has been invested into transferring the probabilistic topic modeling concept from monolingual to multilingual settings. Novel topic models have been designed to work with parallel and comparable texts. We define the concept of multilingual probabilistic topic modeling and present a short high-level overview of the current research and methodology. As a representative example, we thoroughly describe a multilingual probabilistic topic model called bilingual LDA (BiLDA) trained on comparable data in the appendix. In the paper we provide a short overview of cross-lingual applications for which we utilized the model in our research so far.
Download slides: nipsworkshops2012_vulic_topic_modeling_01.pdf (1.4 MB)
Link this pageWould you like to put a link to this lecture on your homepage?
Go ahead! Copy the HTML snippet !