ICCV 2009 Tilke Judd, Krista Ehinger, Fr´edo Durand, Antonio Torralba.

Slides:



Advertisements
Similar presentations
Rapid Object Detection using a Boosted Cascade of Simple Features Paul Viola, Michael Jones Conference on Computer Vision and Pattern Recognition 2001.
Advertisements

What makes an image memorable?
Effect of Opacity of Stimulus in Deployment of Interest in an Interface Sujoy Kumar Chowdhury & Jeremiah D. Still Missouri Western State University Introduction.
Qualifying Exam: Contour Grouping Vida Movahedi Supervisor: James Elder Supervisory Committee: Minas Spetsakis, Jeff Edmonds York University Summer 2009.
A Framework for Photo-Quality Assessment and Enhancement based on Visual Aesthetics Subhabrata Bhattacharya Rahul Sukthankar Mubarak Shah.
Hierarchical Saliency Detection School of Electronic Information Engineering Tianjin University 1 Wang Bingren.
Student: Yao-Sheng Wang Advisor: Prof. Sheng-Jyh Wang ARTICULATED HUMAN DETECTION 1 Department of Electronics Engineering National Chiao Tung University.
Landmark Classification in Large- scale Image Collections Yunpeng Li David J. Crandall Daniel P. Huttenlocher ICCV 2009.
IEEE TCSVT 2011 Wonjun Kim Chanho Jung Changick Kim
Stas Goferman Lihi Zelnik-Manor Ayellet Tal. …
Event prediction CS 590v. Applications Video search Surveillance – Detecting suspicious activities – Illegally parked cars – Abandoned bags Intelligent.
Learning to Compute the Symmetry Plane for Human Faces Jia Wu ACM-BCB '11, August
Small Codes and Large Image Databases for Recognition CVPR 2008 Antonio Torralba, MIT Rob Fergus, NYU Yair Weiss, Hebrew University.
1. Introduction Humanising GrabCut: Learning to segment humans using the Kinect Varun Gulshan, Victor Lempitksy and Andrew Zisserman Dept. of Engineering.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007.
Robust supervised image classifiers by spatial AdaBoost based on robust loss functions Ryuei Nishii and Shinto Eguchi Proc. Of SPIE Vol D-2.
Visual Attention More information in visual field than we can process at a given moment Solutions Shifts of Visual Attention related to eye movements Some.
CVR05 University of California Berkeley 1 Familiar Configuration Enables Figure/Ground Assignment in Natural Scenes Xiaofeng Ren, Charless Fowlkes, Jitendra.
Saliency & attention (P) Lavanya Sharan April 4th, 2011.
A structured learning framework for content- based image indexing and visual Query (Joo-Hwee, Jesse S. Jin) Presentation By: Salman Ahmad (270279)
FACE DETECTION AND RECOGNITION By: Paranjith Singh Lohiya Ravi Babu Lavu.
DVMM Lab, Columbia UniversityVideo Event Recognition Video Event Recognition: Multilevel Pyramid Matching Dong Xu and Shih-Fu Chang Digital Video and Multimedia.
Speaker: Chi-Yu Hsu Advisor: Prof. Jian-Jung Ding Leveraging Stereopsis for Saliency Analysis, CVPR 2012.
Salient Object Detection by Composition
A Generic Approach for Image Classification Based on Decision Tree Ensembles and Local Sub-windows Raphaël Marée, Pierre Geurts, Justus Piater, Louis Wehenkel.
Face Alignment Using Cascaded Boosted Regression Active Shape Models
A Model of Saliency-Based Visual Attention for Rapid Scene Analysis Laurent Itti, Christof Koch, and Ernst Niebur IEEE PAMI, 1998.
Manipulating Attention in Computer Games Matthias Bernhard, Le Zhang, Michael Wimmer Institute of Computer Graphics and Algorithms Vienna University of.
SVCL Automatic detection of object based Region-of-Interest for image compression Sunhyoung Han.
A General Framework for Tracking Multiple People from a Moving Camera
Learning a Fast Emulator of a Binary Decision Process Center for Machine Perception Czech Technical University, Prague ACCV 2007, Tokyo, Japan Jan Šochman.
Assessment of Computational Visual Attention Models on Medical Images Varun Jampani 1, Ujjwal 1, Jayanthi Sivaswamy 1 and Vivek Vaidya 2 1 CVIT, IIIT Hyderabad,
Ranking and Classifying Attractiveness of Photos in Folksonomies Jose San Pedro and Stefan Siersdorfer University of Sheffield, L3S Research Center WWW.
Beauty is Here! Evaluating Aesthetics in Videos Using Multimodal Features and Free Training Data Yanran Wang, Qi Dai, Rui Feng, Yu-Gang Jiang School of.
Stylization and Abstraction of Photographs Doug Decarlo and Anthony Santella.
BAGGING ALGORITHM, ONLINE BOOSTING AND VISION Se – Hoon Park.
UNBIASED LOOK AT DATASET BIAS Antonio Torralba Massachusetts Institute of Technology Alexei A. Efros Carnegie Mellon University CVPR 2011.
Putting Context into Vision Derek Hoiem September 15, 2004.
PSEUDO-RELEVANCE FEEDBACK FOR MULTIMEDIA RETRIEVAL Seo Seok Jun.
Department of Psychology & The Human Computer Interaction Program Vision Sciences Society’s Annual Meeting, Sarasota, FL May 13, 2007 Jeremiah D. Still,
Look Over Here: Attention-Directing Composition of Manga Elements Ying Cao Rynson W.H. Lau Antoni B. Chan SIGGRAPH
Sparse Bayesian Learning for Efficient Visual Tracking O. Williams, A. Blake & R. Cipolloa PAMI, Aug Presented by Yuting Qi Machine Learning Reading.
Region-Based Saliency Detection and Its Application in Object Recognition IEEE TRANSACTIONS ON CIRCUITS AND SYSTEM FOR VIDEO TECHNOLOGY, VOL. 24 NO. 5,
Saliency Aggregation: A Data- driven Approach Long Mai Yuzhen Niu Feng Liu Department of Computer Science, Portland State University Portland, OR,
VIP: Finding Important People in Images Clint Solomon Mathialagan Andrew C. Gallagher Dhruv Batra CVPR
NEUROAESTHETICS IN FASHION: MODELING THE PERCEPTION OF FASHIONABILITY EDGAR SIMO-SERRA SANJA FIDLER FRANCESC MORENO-NOGUER RAQUEL URTASUN CVPR2015.
Image Classification for Automatic Annotation
Stas Goferman Lihi Zelnik-Manor Ayellet Tal Technion.
Spatio-temporal saliency model to predict eye movements in video free viewing Gipsa-lab, Grenoble Département Images et Signal CNRS, UMR 5216 S. Marat,
Towards Total Scene Understanding: Classification, Annotation and Segmentation in an Automatic Framework N 工科所 錢雅馨 2011/01/16 Li-Jia Li, Richard.
POSTER TEMPLATE BY: Background Objectives Psychophysical Experiment Smoothness Features Project Pipeline and outlines The purpose.
Learning video saliency from human gaze using candidate selection CVPR2013 Poster.
SUN Database: Large-scale Scene Recognition from Abbey to Zoo Jianxiong Xiao *James Haysy Krista A. Ehinger Aude Oliva Antonio Torralba Massachusetts Institute.
Photo-Quality Enhancement based on Visual Aesthetics S. Bhattacharya*, R. Sukthankar**, M.Shah* *University of Central Florida, **Intel labs & CMU.
A computational model of stereoscopic 3D visual saliency School of Electronic Information Engineering Tianjin University 1 Wang Bingren.
 Mentor : Prof. Amitabha Mukerjee Learning to Detect Salient Objects Team Members - Avinash Koyya Diwakar Chauhan.
Distinctive Image Features from Scale-Invariant Keypoints Presenter :JIA-HONG,DONG Advisor : Yen- Ting, Chen 1 David G. Lowe International Journal of Computer.
Segmentation of Building Facades using Procedural Shape Priors
REAL-TIME DETECTOR FOR UNUSUAL BEHAVIOR
2. Skin - color filtering.
Krishna Kumar Singh, Yong Jae Lee University of California, Davis
Figure Legend: From: Fixations on low-resolution images
Saliency detection Donghun Yeo CV Lab..
Adri`a Recasens, Aditya Khosla, Carl Vondrick, Antonio Torralba
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
Analysis and classification of images based on focus
Volume 88, Issue 3, Pages (November 2015)
Introduction Task: extracting relational facts from text
Basics of ML Rohan Suri.
Presentation transcript:

ICCV 2009 Tilke Judd, Krista Ehinger, Fr´edo Durand, Antonio Torralba

 Introduction  Database of eye tracking data  Learning a model of saliency  Applications  Conclusion

Bottom-up control of selective attention − stimulus salience (defined by color, contrast and orientation) − Saliency map

 Current saliency models do not accurately predict human fixations.

Top-down control of selective attention − Scene schema guides fixations (more likely to land on meaningful areas) − Task goals guides fixations to land on objects relevant to the task

The first is a large database of eye tracking experiments with labels and analysis Second is a supervised learning model of saliency which combines both bottom-up image based saliency cues and top-down image semantic dependent cues Goal : Predict where users look without the eye tracking hardware.

 Data gathering protocol ◦ 1003 random images from Flickr and LabelMe and recorded eye tracking data from 15 users who free viewed these images. 779 landscape images and 228 portrait images.

 Data gathering protocol Gaze tracking paths and fixation locations are recorded for each viewer

 Data gathering protocol left. Saliency mapright. most salient 20 percent of the image Gaussia n filter

 Analysis of dataset ◦ a strong bias for human fixations to be near the center of the image [19][23]

 Analysis of dataset ◦ the performance of human saliency maps to predict eye fixations Ground truth fixations Saliency map as classifier

 Analysis of dataset ◦ Object of interest and Size of regions of interest

 Features used for machine learning ◦ Low-level features ex: color,orientation,intensity ◦ Mid-level features ex: horizon ◦ High-level features ex: face detector ◦ Center prior : distance to the center

 Training sample selection ◦ 903 training images and 100 testing images ◦ 10 positively labeled pixels randomly from the top 20% salient locations 10 negatively labeled pixels from the bottom 70% salient locations  Training ◦ used the liblinear support vector machine to train a model

 Comparison of saliency maps

 Performance on testing images 1. Outperforms than other model 2. Reaches 88% of the way to human performance 3. not benefit from the huge bias of fixations toward the center 4. the overall performance for the object detector model is low

 Performance on testing samples (the average of the true positive and true negative rates) 1. performs only as well as chance for the other subsets of samples 2. the later model performs more robustly over all subsets of samples 3. people and cars performs better on the subsets with faces

 Using eye tracking data to decide how to render a photograph with differing levels of detail. [4] D. DeCarlo and A. Santella. Stylization and abstraction of photographs. ACM Transactions on Graphics

 Contributions ◦ Developed a largest eye tracking database of natural images and permits large-scale quantitative analysis of fixations points and gaze paths. ◦ Using machine learning to train a bottom-up, top-down model of saliency and outperforms several existing Models.  future work ◦ understanding the impact of framing, cropping and scaling images on fixations.