Making centralized (graph) computation faster, distributed and (at times) better
published: Oct. 16, 2012, recorded: September 2012, views: 76
Report a problem or upload filesIf you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.
Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
I will introduce a generic method for approximate inference in graphical models using graph partitioning. The resulting algorithm is linear time and provides an excellent approximation for the maximum a posteriori assignment (MAP) in a larger class of graphical model including any graph with "polynomial growth" and graph that exclude fixed minors (e.g. planar graphs). In general, the algorithm can be thought of as a "meta" algorithm that can be used to speed up any existing inference algorithm without losing performance. The goal of the talk is to primarily introduce the algorithm and provide insights into why such a simplistic algorithm works. Time permitting, I will also discuss its implication for "modularity clustering" that has been popularly utilized in processing networked data.
Download slides: cyberstat2012_shah_centralized_computation_01.pdf (777.3 KB)
Link this pageWould you like to put a link to this lecture on your homepage?
Go ahead! Copy the HTML snippet !