Focusing Human Attention on the "Right" Visual Data

author: Kristen Grauman, University of Texas at Austin
recorded by: IEEE ICME
published: Sept. 18, 2012,   recorded: July 2012,   views: 8951


Related Open Educational Resources

Related content

Report a problem or upload files

If you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.
Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
Lecture popularity: You need to login to cast your vote.


Widespread visual sensors and unprecedented connectivity have left us awash with visual data‐‐‐from online photo collections, home videos, news footage, medical images, or surveillance feeds. Which images and videos among them warrant human attention? This talk focuses on two problem settings in which this question is critical: supervised learning of object categories, and unsupervised video summarization. In the first setting, the challenge is to sift through candidate training images and select those that, if labeled by a human, would be most informative to the recognition system. In the second, the challenge is to sift through a long‐running video and select only the essential parts needed to summarize it for a human viewer. I will present our recent research addressing these problems, including novel algorithms for large‐scale active learning and egocentric video synopses for wearable cameras. Both domains demonstrate the importance of "semi‐automating" certain computer vision tasks, and suggest exciting new applications for large‐scale visual analysis.

See Also:

Download slides icon Download slides: icme2012_grauman_visual_data_01.pdf (3.8 MB)

Help icon Streaming Video Help

Link this page

Would you like to put a link to this lecture on your homepage?
Go ahead! Copy the HTML snippet !

Write your own review or comment:

make sure you have javascript enabled or clear this field: