Bayesian Decision Theory Case Studies CS479/679 Pattern Recognition Dr. George Bebis.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Bayesian Decision Theory Case Studies
Road-Sign Detection and Recognition Based on Support Vector Machines Saturnino, Sergio et al. Yunjia Man ECG 782 Dr. Brendan.
Face Recognition. Introduction Why we are interested in face recognition? Why we are interested in face recognition? Passport control at terminals in.
Simple Face Detection system Ali Arab Sharif university of tech. Fall 2012.
Learning to estimate human pose with data driven belief propagation Gang Hua, Ming-Hsuan Yang, Ying Wu CVPR 05.
HYBRID-BOOST LEARNING FOR MULTI-POSE FACE DETECTION AND FACIAL EXPRESSION RECOGNITION Hsiuao-Ying ChenChung-Lin Huang Chih-Ming Fu Pattern Recognition,
Robust Object Tracking via Sparsity-based Collaborative Model
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
ICIP 2000, Vancouver, Canada IVML, ECE, NTUA Face Detection: Is it only for Face Recognition?  A few years earlier  Face Detection Face Recognition 
Tracking Objects with Dynamics Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 04/21/15 some slides from Amin Sadeghi, Lana Lazebnik,
Parameter Estimation: Maximum Likelihood Estimation Chapter 3 (Duda et al.) – Sections CS479/679 Pattern Recognition Dr. George Bebis.
Automatic Face Recognition Using Color Based Segmentation and Intelligent Energy Detection Michael Padilla and Zihong Fan Group 16 EE368, Spring
Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007.
Object Detection and Tracking Mike Knowles 11 th January 2005
Effective Gaussian mixture learning for video background subtraction Dar-Shyang Lee, Member, IEEE.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
A Vision-Based System that Detects the Act of Smoking a Cigarette Xiaoran Zheng, University of Nevada-Reno, Dept. of Computer Science Dr. Mubarak Shah,
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
1 Activity and Motion Detection in Videos Longin Jan Latecki and Roland Miezianko, Temple University Dragoljub Pokrajac, Delaware State University Dover,
Computer vision: models, learning and inference Chapter 6 Learning and Inference in Vision.
Flow Based Action Recognition Papers to discuss: The Representation and Recognition of Action Using Temporal Templates (Bobbick & Davis 2001) Recognizing.
ECSE 6610 Pattern Recognition Professor Qiang Ji Spring, 2011.
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
1 Mean shift and feature selection ECE 738 course project Zhaozheng Yin Spring 2005 Note: Figures and ideas are copyrighted by original authors.
Human-Computer Interaction Human-Computer Interaction Tracking Hanyang University Jong-Il Park.
EE 492 ENGINEERING PROJECT LIP TRACKING Yusuf Ziya Işık & Ashat Turlibayev Yusuf Ziya Işık & Ashat Turlibayev Advisor: Prof. Dr. Bülent Sankur Advisor:
An Information Fusion Approach for Multiview Feature Tracking Esra Ataer-Cansizoglu and Margrit Betke ) Image and.
Object Recognition in Images Slides originally created by Bernd Heisele.
Face Recognition: An Introduction
Stable Multi-Target Tracking in Real-Time Surveillance Video
Face Detection Ying Wu Electrical and Computer Engineering Northwestern University, Evanston, IL
Action and Gait Recognition From Recovered 3-D Human Joints IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS— PART B: CYBERNETICS, VOL. 40, NO. 4, AUGUST.
Efficient Visual Object Tracking with Online Nearest Neighbor Classifier Many slides adapt from Steve Gu.
Expectation-Maximization (EM) Case Studies
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
Boosted Particle Filter: Multitarget Detection and Tracking Fayin Li.
A Reliable Skin Detection Using Dempster-Shafer Theory of Evidence
 Present by 陳群元.  Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental.
Irfan Ullah Department of Information and Communication Engineering Myongji university, Yongin, South Korea Copyright © solarlits.com.
Face Detection and Head Tracking Ying Wu Electrical Engineering & Computer Science Northwestern University, Evanston, IL
Cell Segmentation in Microscopy Imagery Using a Bag of Local Bayesian Classifiers Zhaozheng Yin RI/CMU, Fall 2009.
Learning and Removing Cast Shadows through a Multidistribution Approach Nicolas Martel-Brisson, Andre Zaccarin IEEE TRANSACTIONS ON PATTERN ANALYSIS AND.
Face Detection 蔡宇軒.
Gaussian Mixture Model classification of Multi-Color Fluorescence In Situ Hybridization (M-FISH) Images Amin Fazel 2006 Department of Computer Science.
Motion Estimation of Moving Foreground Objects Pierre Ponce ee392j Winter March 10, 2004.
CS479/679 Pattern Recognition Dr. George Bebis
Computer vision: models, learning and inference
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
A Forest of Sensors: Using adaptive tracking to classify and monitor activities in a site Eric Grimson AI Lab, Massachusetts Institute of Technology
Tracking Objects with Dynamics
Face Recognition and Feature Subspaces
Real-Time Human Pose Recognition in Parts from Single Depth Image
Dynamical Statistical Shape Priors for Level Set Based Tracking
Chapter 3: Maximum-Likelihood and Bayesian Parameter Estimation (part 2)
Lecture 25: Introduction to Recognition
Eric Grimson, Chris Stauffer,
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
George Bebis and Wenjing Li Computer Vision Laboratory
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
Announcements Project 2 artifacts Project 3 due Thursday night
Announcements Project 4 out today Project 2 winners help session today
EE 492 ENGINEERING PROJECT
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
EM Algorithm and its Applications
Chapter 3: Maximum-Likelihood and Bayesian Parameter Estimation (part 2)
Presentation transcript:

Bayesian Decision Theory Case Studies CS479/679 Pattern Recognition Dr. George Bebis

Case Study I A. Madabhushi and J. Aggarwal, A bayesian approach to human activity recognition, 2nd International Workshop on Visual Surveillance, pp , June 1999.A bayesian approach to human activity recognition

Human activity recognition Recognize human actions using visual information. – Useful for monitoring of human activity in department stores, airports, high-security buildings etc. Building systems that can recognize any type of action is a difficult and challenging problem.

Goal (this paper) Build a system that is capable of recognizing the following 10 (ten) actions, from a frontal or lateral view: sitting down standing up bending down getting up hugging squatting rising from a squatting position bending sideways falling backward walking

Motivation People sit, stand, walk, bend down, and get up in a more or less similar fashion. Human actions can be recognized by tracking various body parts.

Proposed Approach Use head motion trajectory – The head of a person moves in a characteristic fashion during these actions. Recognition is formulated as Bayesian classification using the movement of the head over consecutive frames.

Strengths and Weaknesses Strengths – The system can recognize actions where the gait of the subject in the test sequence differs considerably from the training sequences. – Also, it can recognize actions for people of varying physical appearance (i.e., tall, short, fat, thin etc.). Limitations – Only actions in the frontal or lateral view can be recognized successfully by this system. – Non-realistic assumptions.

Main Steps input output

Action Representation Estimate the centroid of the head in each frame: Find the absolute differences in successive frames: ||| |,

Head Detection and Tracking Accurate head detection and tracking are crucial. In this paper, the centroid of the head was tracked manually from frame to frame.

Bayesian Formulation Given an input sequence, the posterior probabilities are computed for each action (2 x 10 = 20) using the Bayes rule:

Probability Density Estimation Feature vectors X and Y are assumed to be independent (valid?), following a multi-variate Gaussian distribution:

Probability Density Estimation (cont’d) The sample covariance matrices are used to estimate Σ X and Σ Y : Two distributions are estimated for each action, corresponding to the frontal and lateral views. ΣXΣX ΣYΣY

Action Classification Given an input sequence, the posterior probability is computed for each action. The unknown action is classified based on the most likely action:

Discriminating Similar Actions In some actions, the head moves in a similar fashion, making it difficult to distinguish these actions from one another; i.e.,: (1) The head moves downward without much sideward deviation in the following actions: * squatting * sitting down * bending down

Discriminating Similar Actions (cont’d) (2) The head moves upward without much sideward deviation in the following actions: * standing up * rising * getting up Several heuristics are used to distinguish among similar actions: – e.g., when bending down, the head goes much lower than when sitting down.

Training A fixed CCD camera working at 2 frames per second was used to obtain the training data. – People of diverse physical appearance were used to model the actions. – Subjects were asked to perform the actions at a comfortable pace. – 38 sequences were taken of each person performing all the actions in both the frontal and lateral views.

Training (cont’d) Assumptions – It was found that each action can be completed within 10 frames. – Only the first 10 frames from each sequence were used for training (i.e., 5 seconds)

Testing 39 sequences were used for testing Only the first 10 frames from each sequence were used for testing (i.e., 5 seconds) Of the 8 sequences classified incorrectly, 6 were assigned to the correct action but to the wrong view.

Practical Issues How would you find the first and last frames of an action in general (segmentation)? Is the system robust to recognizing an action performed at various speeds or from incomplete sequences (i.e., assuming that several frames are missing)? Current system is unable to recognize actions from different viewpoints.

Extension J. Usabiaga, G. Bebis, A. Erol, Mircea Nicolescu, and Monica Nicolescu, "Recognizing Simple Human Actions Using 3D Head Trajectories", Computational Intelligence, vol. 23, no. 4, pp , 2007.Recognizing Simple Human Actions Using 3D Head Trajectories

Case Study II J. Yang and A. Waibel, A Real-time Face Tracker, Proceedings of WACV'96, 1996.A Real-time Face Tracker

Overview Build a system that can detect and track a person’s face while the person moves freely in some environment. – Useful in a number of applications such as video conference, visual surveillance, face recognition, etc. Key Idea: Use a skin color model to detect faces in an image.

Why Using Skin Color? Traditional systems for face detection use template matching or facial features. – Not very robust and time consuming. Using skin-color leads to faster and more robust face detection.

Main Steps (1) Detect human faces in various using a skin-color model. (2) Track face of interest by controlling the camera position and zoom. (3) Adapt skin-color model parameters based on individual appearance and lighting conditions.

Main System Components A probabilistic model to characterize skin- color distributions of human faces. A motion model to estimate human motion and to predict search window in the next frame. A camera model to predict camera motion (i.e., camera’s response was much slower than frame rate).

Search Window

Challenges Modeling Skin Color Skin color is influenced by many factors: – Skin color varies from person to person. – Skin color can be affected by ambient light, motion etc. – Different cameras produce significantly different color values (i.e., even for the same person under the same lighting conditions).

RGB vs Chromatic Color Space RGB is not the best color representation for characterizing skin-color (i.e., it represents not only color but also brightness). Represent skin-color in the chromatic space which is defined as follows: (note: the normalized blue value is redundant since r + g + b = 1)

Skin-Color Clustering Skin colors do not fall randomly in the chromatic color space but actually form clusters.

Skin-Color Clustering (cont’d) Skin-colors of different people are also clustered in chromatic color space – i.e., they differ more in brightness than in color. (skin-color distribution of 40 people - different races)

Skin-Color Modeling Experiments (i.e., under different lighting conditions and persons) have shown that the skin-color distribution has a rather regular shape. Idea: represent skin-color distribution using a 2D Gaussian distribution with mean μ and covariance Σ: Examples:

Parameter Estimation Collect skin-color regions from a set of face images. Estimate the mean and covariance using the sample mean and sample covariance:

Face detection using skin-color Each pixel x in the input image is converted into the chromatic color space and compared with the distribution of the skin-color model.

Example Note: in general, we can model the non-skin-color distribution too and compute the max posterior probability using the Bayes rule (i.e., two-class classification: skin-color vs non-skin-color)

Dealing with skin-color-like objects It is impossible in general to detect only faces simply from the result of color matching. – e.g., background may contain skin colors

Dealing with skin-color-like objects (cont’d) Additional information should be used for rejecting false positives (e.g., geometric features, motion etc.)

Skin-color model adaptation If a person is moving, the apparent skin colors change as the person’s position relative to the camera or light changes. Idea: adapt model parameters (μ,Σ) to handle these changes.

Skin-color model adaptation (cont’d) The weighting factors a i, b i, c i determine how much past parameters will influence current parameters. N determines how long the past parameters will influence the current parameters.

System initialization Automatic mode – A general skin-color model is used to identify skin- color regions. – Motion and shape information is used to reject non-face regions. – The largest face region is selected (i.e., face closest to the camera). – Skin-color model is adapted to the face being tracked.

System initialization (cont’d) Interactive mode – The user selects a point on the face of interest using the mouse. – The tracker searches around the point to find the face using a general skin-color model. – Skin-color model is adapted to the face being tracked.

Detection Results