Efficient algorithms for estimating multi-view mixture models

author: Daniel Hsu, Microsoft Research New England, Microsoft Research
published: Jan. 16, 2013,   recorded: December 2012,   views: 2989

Slides

Related Open Educational Resources

Related content

Report a problem or upload files

If you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.
Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
Lecture popularity: You need to login to cast your vote.
  Bibliography

Description

Mixture models are a staple in machine learning and applied statistics for treating data taken from multiple sub-populations. For many classes of mixture models, parameter estimation is computationally and/or information-theoretically hard in general. However, much progress has been made over the past decade or so to overcome these hardness barriers by focusing on sub-classes that rule out the intractable cases.

One very powerful and general sub-class is the multi-view setting, where one can take advantage of several non-redundant sources of information to help distinguish different sub-populations. In this talk, I'll describe a general technique that is applicable even in semi-parametric settings, where one may not have a parametric model for individual mixture components. This technique also yields a number of new unsupervised learning results for well-studied problems, as well as very practical and scalable learning algorithms.

See Also:

Download slides icon Download slides: nipsworkshops2012_hsu_algorithms_01.pdf (3.2┬áMB)


Help icon Streaming Video Help

Link this page

Would you like to put a link to this lecture on your homepage?
Go ahead! Copy the HTML snippet !

Write your own review or comment:

make sure you have javascript enabled or clear this field: