Focusing Human Attention on the "Right" Visual Data
Report a problem or upload filesIf you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.
Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
Widespread visual sensors and unprecedented connectivity have left us awash with visual data‐‐‐from online photo collections, home videos, news footage, medical images, or surveillance feeds. Which images and videos among them warrant human attention? This talk focuses on two problem settings in which this question is critical: supervised learning of object categories, and unsupervised video summarization. In the first setting, the challenge is to sift through candidate training images and select those that, if labeled by a human, would be most informative to the recognition system. In the second, the challenge is to sift through a long‐running video and select only the essential parts needed to summarize it for a human viewer. I will present our recent research addressing these problems, including novel algorithms for large‐scale active learning and egocentric video synopses for wearable cameras. Both domains demonstrate the importance of "semi‐automating" certain computer vision tasks, and suggest exciting new applications for large‐scale visual analysis.
Link this pageWould you like to put a link to this lecture on your homepage?
Go ahead! Copy the HTML snippet !