CVPR2013 Poster Modeling Actions through State Changes.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Complex Networks for Representation and Characterization of Images For CS790g Project Bingdong Li 9/23/2009.
Zhimin CaoThe Chinese University of Hong Kong Qi YinITCS, Tsinghua University Xiaoou TangShenzhen Institutes of Advanced Technology Chinese Academy of.
RGB-D object recognition and localization with clutter and occlusions Federico Tombari, Samuele Salti, Luigi Di Stefano Computer Vision Lab – University.
Foreground Focus: Finding Meaningful Features in Unlabeled Images Yong Jae Lee and Kristen Grauman University of Texas at Austin.
Change Detection C. Stauffer and W.E.L. Grimson, “Learning patterns of activity using real time tracking,” IEEE Trans. On PAMI, 22(8): , Aug 2000.
Human Identity Recognition in Aerial Images Omar Oreifej Ramin Mehran Mubarak Shah CVPR 2010, June Computer Vision Lab of UCF.
Computer Vision Lecture 18: Object Recognition II
Patch to the Future: Unsupervised Visual Prediction
Identifying Image Spam Authorship with a Variable Bin-width Histogram-based Projective Clustering Song Gao, Chengcui Zhang, Wei Bang Chen Department of.
Activity Recognition Aneeq Zia. Agenda What is activity recognition Typical methods used for action recognition “Evaluation of local spatio-temporal features.
Real-Time Human Pose Recognition in Parts from Single Depth Images Presented by: Mohammad A. Gowayyed.
Watching Unlabeled Video Helps Learn New Human Actions from Very Few Labeled Snapshots Chao-Yeh Chen and Kristen Grauman University of Texas at Austin.
Large-Scale Object Recognition with Weak Supervision
Sketch Tokens: A Learned Mid-level Representation for Contour and Object Detection CVPR2013 POSTER.
Real-time Embedded Face Recognition for Smart Home Fei Zuo, Student Member, IEEE, Peter H. N. de With, Senior Member, IEEE.
A Bayesian algorithm for tracking multiple moving objects in outdoor surveillance video Department of Electrical Engineering and Computer Science The University.
CS335 Principles of Multimedia Systems Content Based Media Retrieval Hao Jiang Computer Science Department Boston College Dec. 4, 2007.
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman University of Oxford ICCV 2003.
CS 223B Assignment 1 Help Session Dan Maynes-Aminzade.
MULTIPLE MOVING OBJECTS TRACKING FOR VIDEO SURVEILLANCE SYSTEMS.
Introduction to Object Tracking Presented by Youyou Wang CS643 Texas A&M University.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
1 Chapter 21 Machine Vision. 2 Chapter 21 Contents (1) l Human Vision l Image Processing l Edge Detection l Convolution and the Canny Edge Detector l.
 For many years human being has been trying to recreate the complex mechanisms that human body forms & to copy or imitate human systems  As a result.
1 Activity and Motion Detection in Videos Longin Jan Latecki and Roland Miezianko, Temple University Dragoljub Pokrajac, Delaware State University Dover,
Wang, Z., et al. Presented by: Kayla Henneman October 27, 2014 WHO IS HERE: LOCATION AWARE FACE RECOGNITION.
A Scale and Rotation Invariant Approach to Tracking Human Body Part Regions in Videos Yihang BoHao Jiang Institute of Automation, CAS Boston College.
Computer vision.
Flow Based Action Recognition Papers to discuss: The Representation and Recognition of Action Using Temporal Templates (Bobbick & Davis 2001) Recognizing.
TP15 - Tracking Computer Vision, FCUP, 2013 Miguel Coimbra Slides by Prof. Kristen Grauman.
Unsupervised Object Segmentation with a Hybrid Graph Model (HGM) Reporter: 鄭綱 (6/14)
Watch, Listen and Learn Sonal Gupta, Joohyun Kim, Kristen Grauman and Raymond Mooney -Pratiksha Shah.
BACKGROUND LEARNING AND LETTER DETECTION USING TEXTURE WITH PRINCIPAL COMPONENT ANALYSIS (PCA) CIS 601 PROJECT SUMIT BASU FALL 2004.
Marcin Marszałek, Ivan Laptev, Cordelia Schmid Computer Vision and Pattern Recognition, CVPR Actions in Context.
Interactive Discovery and Semantic Labeling of Patterns in Spatial Data Thomas Funkhouser, Adam Finkelstein, David Blei, and Christiane Fellbaum Princeton.
Detecting Pedestrians Using Patterns of Motion and Appearance Paul Viola Microsoft Research Irfan Ullah Dept. of Info. and Comm. Engr. Myongji University.
Building local part models for category-level recognition C. Schmid, INRIA Grenoble Joint work with G. Dorko, S. Lazebnik, J. Ponce.
Texture. Texture is an innate property of all surfaces (clouds, trees, bricks, hair etc…). It refers to visual patterns of homogeneity and does not result.
#MOTION ESTIMATION AND OCCLUSION DETECTION #BLURRED VIDEO WITH LAYERS
Face Recognition: An Introduction
An Efficient Search Strategy for Block Motion Estimation Using Image Features Digital Video Processing 1 Term Project Feng Li Michael Su Xiaofeng Fan.
Dynamic Captioning: Video Accessibility Enhancement for Hearing Impairment Richang Hong, Meng Wang, Mengdi Xuy Shuicheng Yany and Tat-Seng Chua School.
Event retrieval in large video collections with circulant temporal encoding CVPR 2013 Oral.
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
Image-Based Segmentation of Indoor Corridor Floors for a Mobile Robot
Kylie Gorman WEEK 1-2 REVIEW. CONVERTING AN IMAGE FROM RGB TO HSV AND DISPLAY CHANNELS.
Image-Based Segmentation of Indoor Corridor Floors for a Mobile Robot Yinxiao Li and Stanley T. Birchfield The Holcombe Department of Electrical and Computer.
Data Mining, ICDM '08. Eighth IEEE International Conference on Duy-Dinh Le National Institute of Informatics Hitotsubashi, Chiyoda-ku Tokyo,
Text From Corners: A Novel Approach to Detect Text and Caption in Videos Xu Zhao, Kai-Hsiang Lin, Yun Fu, Member, IEEE, Yuxiao Hu, Member, IEEE, Yuncai.
Multi-view Traffic Sign Detection, Recognition and 3D Localisation Radu Timofte, Karel Zimmermann, and Luc Van Gool.
SUMMERY 1. VOLUMETRIC FEATURES FOR EVENT DETECTION IN VIDEO correlate spatio-temporal shapes to video clips that have been automatically segmented we.
Surface Defect Inspection: an Artificial Immune Approach Dr. Hong Zheng and Dr. Saeid Nahavandi School of Engineering and Technology.
Motion tracking TEAM D, Project 11: Laura Gui - Timisoara Calin Garboni - Timisoara Peter Horvath - Szeged Peter Kovacs - Debrecen.
Week 3 Emily Hand UNR. Online Multiple Instance Learning The goal of MIL is to classify unseen bags, instances, by using the labeled bags as training.
南台科技大學 資訊工程系 Region partition and feature matching based color recognition of tongue image 指導教授:李育強 報告者 :楊智雁 日期 : 2010/04/19 Pattern Recognition Letters,
Another Example: Circle Detection
Finding Things: Image Parsing with Regions and Per-Exemplar Detectors
Nonparametric Semantic Segmentation
Video Google: Text Retrieval Approach to Object Matching in Videos
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
Tremor Detection Using Motion Filtering and SVM Bilge Soran, Jenq-Neng Hwang, Linda Shapiro, ICPR, /16/2018.
Introduction Computer vision is the analysis of digital images
Outline Neural networks - reviewed Texture modeling
Brief Review of Recognition + Context
Coupled Horn-Schunck and Lukas-Kanade for image processing
Video Google: Text Retrieval Approach to Object Matching in Videos
Introduction Computer vision is the analysis of digital images
Multi-Information Based GCPs Selection Method
Support vector machine-based text detection in digital video
Presentation transcript:

CVPR2013 Poster Modeling Actions through State Changes

Outline 1.Introduction 2.Method 3.Results 4.Dicussion

1. Introduction What makes an action (e.g. “open the jar”) identifiable? How can we tell if such an action is performed?.

1. Introduction We are interested in two kinds of changes 1. changes in the state of objects 2. changes the transformation of stuff we adopt egocentric paradigm for recognizing daily activities and actions. we propose a weakly supervised method for learning the object and material states that are necessary for recognizing daily actions.

2. Method 2-1 Discovering State-Specific Regions 2-2 States as Action Requirements 2-3 Modeling Actions through State Changes 2-4 Activity Segmentation

2-1 Discovering State-Specific Regions two assumptions (1) an object state or material does not change unless an action is performed (2) an object state or material change is associated with an action only if it consistently occurs at all instances of that action.

2-1 Discovering State-Specific Regions 1. we identify regions that either appear or disappear as a result of the execution of each action instance. 2. we prune changes that are not consistently associated with an action. 3. we learn a detector for each group of discovered state- specific regions.

Change Detection we find the regions that either appear or disappear we sample a few frames from its beginning and a few frames from its end. we match their pixels using large displacement optical flow [3]. Then for each pair of matched pixels, we compute change based on their color difference

We compare each beginning (ending) image to multiple ending (beginning) images. We set the amount of change to the minimum amount computed among all the comparisons.

Consistent Regions We only keep the subset of those regions that consistently occur across the instances of an action type.

State-Specific Region Detectors Purpose: learn for cluster We describe each region Color:128 Texture:256 Shape:16

2-2 State as Action Requirements Environment state (before/after) Two Criteria Two way Regions exist/absent ? Moved by hands ? Response vector Chose highest score > Threshold ->1 Other -> 0 Foreground segmentation method

2-3 Modeling Action through State Changes

2-4 Activity Segmentation takes all the action detection scores as input and infers the frames that are assigned to each action in that video Train two state detectors State detectors are trained using linear SVM Result in two |A| x T matrices

2-4 Activity Segmentation

3.Result 3-1 Action Recognition 3-2 Activity Segmentation

3-1 Action Recognition Use GTEA Datebase STIP SIFT Previous work Our method

3-2 Activity Segmentation

4.Dicussion building a taxonomy of possible states of objects and materials. A potential challenge would be modeling the changes that do not correspond to observable visual patterns.