Modeling the Shape of a Scene: Seeing the trees as a forest Scene Understanding Seminar 20090203.

Slides:



Advertisements
Similar presentations
The blue and green colors are actually the same.
Advertisements

November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
LARGE-SCALE IMAGE PARSING Joseph Tighe and Svetlana Lazebnik University of North Carolina at Chapel Hill road building car sky.
Internet Vision - Lecture 3 Tamara Berg Sept 10. New Lecture Time Mondays 10:00am-12:30pm in 2311 Monday (9/15) we will have a general Computer Vision.
Ghunhui Gu, Joseph J. Lim, Pablo Arbeláez, Jitendra Malik University of California at Berkeley Berkeley, CA
São Paulo Advanced School of Computing (SP-ASC’10). São Paulo, Brazil, July 12-17, 2010 Looking at People Using Partial Least Squares William Robson Schwartz.
Event prediction CS 590v. Applications Video search Surveillance – Detecting suspicious activities – Illegally parked cars – Abandoned bags Intelligent.
EE 7730 Image Segmentation.
Fast and Compact Retrieval Methods in Computer Vision Part II A. Torralba, R. Fergus and Y. Weiss. Small Codes and Large Image Databases for Recognition.
Statistics of Natural Image Categories Antonio Torralba and Aude Oliva. Network: Computation in Neural Systems, 14(2003) Jonathan Huang
LARGE-SCALE NONPARAMETRIC IMAGE PARSING Joseph Tighe and Svetlana Lazebnik University of North Carolina at Chapel Hill CVPR 2011Workshop on Large-Scale.
Object Recognition: History and Overview Slides adapted from Fei-Fei Li, Rob Fergus, Antonio Torralba, and Jean Ponce.
Regionalized Variables take on values according to spatial location. Given: Where: A “structural” coarse scale forcing or trend A random” Local spatial.
Christian Siagian Laurent Itti Univ. Southern California, CA, USA
Lecture 14: Classification Thursday 18 February 2010 Reading: Ch – 7.19 Last lecture: Spectral Mixture Analysis.
Introduction to Mapping Sciences: Lecture #5 (Form and Structure) Form and Structure Describing primary and secondary spatial elements Explanation of spatial.
Image Classification To automatically categorize all pixels in an image into land cover classes or themes.
Sufficient Dimensionality Reduction with Irrelevance Statistics Amir Globerson 1 Gal Chechik 2 Naftali Tishby 1 1 Center for Neural Computation and School.
Multiple Object Class Detection with a Generative Model K. Mikolajczyk, B. Leibe and B. Schiele Carolina Galleguillos.
Large Scale Recognition and Retrieval. What does the world look like? High level image statistics Object Recognition for large-scale search Focus on scaling.
Representing Acoustic Information
Overview Introduction to local features
Review: Intro to recognition Recognition tasks Machine learning approach: training, testing, generalization Example classifiers Nearest neighbor Linear.
Introduction --Classification Shape ContourRegion Structural Syntactic Graph Tree Model-driven Data-driven Perimeter Compactness Eccentricity.
Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem
Distinctive Image Features from Scale-Invariant Keypoints By David G. Lowe, University of British Columbia Presented by: Tim Havinga, Joël van Neerbos.
CSE 185 Introduction to Computer Vision Pattern Recognition.
Internet-scale Imagery for Graphics and Vision James Hays cs195g Computational Photography Brown University, Spring 2010.
Digital Image Processing, 2nd ed. © 2002 R. C. Gonzalez & R. E. Woods Chapter 11 Representation & Description Chapter 11 Representation.
General Tensor Discriminant Analysis and Gabor Features for Gait Recognition by D. Tao, X. Li, and J. Maybank, TPAMI 2007 Presented by Iulian Pruteanu.
Computational Photography Tamara Berg Features. Representing Images Keep all the pixels! Pros? Cons?
Why Categorize in Computer Vision ?. Why Use Categories? People love categories!
Recovering Surface Layout from a Single Image D. Hoiem, A.A. Efros, M. Hebert Robotics Institute, CMU Presenter: Derek Hoiem CS 598, Spring 2009 Jan 29,
SVM-KNN Discriminative Nearest Neighbor Classification for Visual Category Recognition Hao Zhang, Alex Berg, Michael Maire, Jitendra Malik.
Classifying Images with Visual/Textual Cues By Steven Kappes and Yan Cao.
Image Categorization Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/15/11.
Features-based Object Recognition P. Moreels, P. Perona California Institute of Technology.
Digital Image Processing, 2nd ed. © 2002 R. C. Gonzalez & R. E. Woods Representation & Description.
Putting Context into Vision Derek Hoiem September 15, 2004.
Introduction --Classification Shape ContourRegion Structural Syntactic Graph Tree Model-driven Data-driven Perimeter Compactness Eccentricity.
Pulkit Agrawal Y7322 BVV Sri Raj Dutt Y7110 Sushobhan Nayak Y7460.
Implementing GIST on the GPU. Refrence Original Work  Aude Oliva, Antonio Torralba  Modeling the shape of the scene: a holistic representation of the.
Discrete Fourier Transform in 2D – Chapter 14. Discrete Fourier Transform – 1D Forward Inverse M is the length (number of discrete samples)
Epitomic Location Recognition A generative approach for location recognition K. Ni, A. Kannan, A. Criminisi and J. Winn In proc. CVPR Anchorage,
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
Colour and Texture. Extract 3-D information Using Vision Extract 3-D information for performing certain tasks such as manipulation, navigation, and recognition.
Extracting Simple Verb Frames from Images Toward Holistic Scene Understanding Prof. Daphne Koller Research Group Stanford University Geremy Heitz DARPA.
Line Matching Jonghee Park GIST CV-Lab..  Lines –Fundamental feature in many computer vision fields 3D reconstruction, SLAM, motion estimation –Useful.
Using the Forest to see the Trees: A computational model relating features, objects and scenes Antonio Torralba CSAIL-MIT Joint work with Aude Oliva, Kevin.
Max-Margin Training of Upstream Scene Understanding Models Jun Zhu Carnegie Mellon University Joint work with Li-Jia Li *, Li Fei-Fei *, and Eric P. Xing.
1 Review and Summary We have covered a LOT of material, spending more time and more detail on 2D image segmentation and analysis, but hopefully giving.
The SUN Database Slides by Jennifer Baulier. What is the SUN database? Scene Understanding Database Scene = a place humans could act within Full database.
Finding Clusters within a Class to Improve Classification Accuracy Literature Survey Yong Jae Lee 3/6/08.
Biologically Inspired Vision-based Indoor Localization Zhihao Li, Ming Yang
Intrinsic Data Geometry from a Training Set
Data Driven Attributes for Action Detection
Quantifying Scale and Pattern Lecture 7 February 15, 2005
Nonparametric Semantic Segmentation
Map of the Great Divide Basin, Wyoming, created using a neural network and used to find likely fossil beds See:
Incorporating Ancillary Data for Classification
Context-based vision system for place and object recognition
Object detection as supervised classification
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
Lecture 25: Introduction to Recognition
CS 1674: Intro to Computer Vision Scene Recognition
Brief Review of Recognition + Context
Unsupervised Classification
CS4670: Intro to Computer Vision
Some slides: courtesy of David Jacobs
Feature Extraction (I)
Presentation transcript:

Modeling the Shape of a Scene: Seeing the trees as a forest Scene Understanding Seminar

Scene recognition Images –“objects”: 1-2 meters –“environments”: > 5 meters This paper –Scene representation –Scene statistics

Scene recognition Scenes vs. collections of objects Object information may be ignored –Fast categorization –Low spatial frequencies –Change blindness, inattention blindness Scene category provides context for images Need holistic representation of a scene

Holistic scene representation Finding a low- dimensional “scene space” Clustering by humans –Split images into groups –ignore objects, categories

Spatial envelope properties Naturalness –natural vs. man-made environments

Spatial envelope properties Openness –decreases as number of boundary elements increases

Spatial envelope properties Roughness –size of elements at each spatial scale, related to fractal dimension

Spatial envelope properties Expansion (man-made environments) –depth gradient of the space

Spatial envelope properties Ruggedness (natural environments) –deviation of ground relative to horizon

Scene statistics DFT (energy spectrum) –throw out phase function (represents local properties) Windowed DFT (spectrogram) –Coarse local information –8x8 grid for these results

Scene statistics Dimensionality reduction via PCA

Scene classification from statistics Different scene categories have different spectral signatures –Amplitude captures roughness –Orientation captures dominant edges

Scene classification from statistics Open environments have non-stationary second-order statistics –support surfaces Closed environments exhibit stationary second-order statistics a) man-made open environments b) urban vertically structured environments c) perspective views of streets d) far view of city-center buildings e) close-up views of urban structures f) natural open environments g) natural closed environments h) mountainous landscapes i) enclosed forests j) close-up views of non-textured scenes

Learning the spatial envelope Use linear regression to learn –DST (discriminant spectral template) –WDST (windowed discriminant spectral template) Relate spectral representation to each spatial envelope feature

Learning the spatial envelope Primacy of Man-made vs. Natural distinction –Linear Discriminant analysis –93.5% correct classification Role of spatial information –WDST not much better than DST –Loschky, et al., scene inversion

Learning the spatial envelope Other properties calculated separately for natural, man-made environments

Spatial envelope and categories Choose random scene and seven neighbors in scene space If >= 4 neighbors have same semantic category, image is “correctly recognized” –WDST: 92% –DST: 86%

Applications Depth Estimation (Torralba & Oliva)

Applications Image Classification (Bosch & Zisserman) –Gist features used in object descriptions

Applications Scene Completion (Hayes & Efros) –Use gist to find find possible matches

Applications Robot Navigation (Siagian & Itti) –Different scene model used to recognize familiar environments

Questions?