Microsoft Kinect How does a machine infer body position?

Slides:



Advertisements
Similar presentations
Miroslav Hlaváč Martin Kozák Fish position determination in 3D space by stereo vision.
Advertisements

For Internal Use Only. © CT T IN EM. All rights reserved. 3D Reconstruction Using Aerial Images A Dense Structure from Motion pipeline Ramakrishna Vedantam.
Víctor Ponce Miguel Reyes Xavier Baró Mario Gorga Sergio Escalera Two-level GMM Clustering of Human Poses for Automatic Human Behavior Analysis Departament.
Multimedia Specification Design and Production 2012 / Semester 1 / week 6 Lecturer: Dr. Nikos Gazepidis
11/06/14 How the Kinect Works Computational Photography Derek Hoiem, University of Illinois Photo frame-grabbed from:
Robust Part-Based Hand Gesture Recognition Using Kinect Sensor
Po-Hsiang Chen Advisor: Sheng-Jyh Wang 2/13/2012.
By : Adham Suwan Mohammed Zaza Ahmed Mafarjeh. Achieving Security through Kinect using Skeleton Analysis (ASKSA)
A Modified EM Algorithm for Hand Gesture Segmentation in RGB-D Data 2014 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE) July 6-11, 2014, Beijing,
Nathan Johnson1 Background Subtraction Various Methods for Different Inputs.
1.Introduction 2.Article [1] Real Time Motion Capture Using a Single TOF Camera (2010) 3.Article [2] Real Time Human Pose Recognition In Parts Using a.
Real-Time Human Pose Recognition in Parts from Single Depth Images Presented by: Mohammad A. Gowayyed.
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
Kinect Case Study CSE P 576 Larry Zitnick
09/18/03CS679 - Fall Copyright Univ. of Wisconsin Last Time Bump Mapping Multi-pass algorithms.
12/01/11 How the Kinect Works Computational Photography Derek Hoiem, University of Illinois Photo frame-grabbed from:
Computational Photography lecture 19 – How the Kinect 1 works? CS 590 Spring 2014 Prof. Alex Berg (Credits to many other folks on individual slides)
(CONTROLLER-FREE GAMING
A Brief Overview of Computer Vision Jinxiang Chai.
Professor : Yih-Ran Sheu Student’s name : Nguyen Van Binh Student ID: MA02B203 Kinect camera 1 Southern Taiwan University Department of Electrical Engineering.
Introduction Kinect for Xbox 360, referred to as Kinect, is developed by Microsoft, used in Xbox 360 video game console and Windows PCs peripheral equipment.
Computer Graphics Computer Graphics is everywhere: Visual system is most important sense: High bandwidth Natural communication Fast developments in Hardware.
3D Fingertip and Palm Tracking in Depth Image Sequences
Zhengyou Zhang Microsoft Research Digital Object Identifier: /MMUL Publication Year: 2012, Page(s): Professor: Yih-Ran Sheu Student.
INTERACTING WITH SIMULATION ENVIRONMENTS THROUGH THE KINECT Fayez Alazmi Supervisor: Dr. Brett Wilkinson Flinders University Image 1Image 2 Source : 1.
KinectFusion : Real-Time Dense Surface Mapping and Tracking IEEE International Symposium on Mixed and Augmented Reality 2011 Science and Technology Proceedings.
1. Introduction Motion Segmentation The Affine Motion Model Contour Extraction & Shape Estimation Recursive Shape Estimation & Motion Estimation Occlusion.
Human Gesture Recognition Using Kinect Camera Presented by Carolina Vettorazzo and Diego Santo Orasa Patsadu, Chakarida Nukoolkit and Bunthit Watanapa.
Probabilistic Context Free Grammars for Representing Action Song Mao November 14, 2000.
A General Framework for Tracking Multiple People from a Moving Camera
Page 1 | Microsoft Work With Color Data Kinect for Windows Video Courses Jan 2013.
A Method for Hand Gesture Recognition Jaya Shukla Department of Computer Science Shiv Nadar University Gautam Budh Nagar, India Ashutosh Dwivedi.
Shape from Stereo  Disparity between two images  Photogrammetry  Finding Corresponding Points Correlation based methods Feature based methods.
December 4, 2014Computer Vision Lecture 22: Depth 1 Stereo Vision Comparing the similar triangles PMC l and p l LC l, we get: Similarly, for PNC r and.
S ENSORS U SED I N G AMES By Wusqa Waqar. What are sensors and how are they used in games? A sensor is a converter that measures a physical quantity and.
DIEGO AGUIRRE COMPUTER VISION INTRODUCTION 1. QUESTION What is Computer Vision? 2.
Online Kinect Handwritten Digit Recognition Based on Dynamic Time Warping and Support Vector Machine Journal of Information & Computational Science, 2015.
Human pose recognition from depth image MS Research Cambridge.
Epitomic Location Recognition A generative approach for location recognition K. Ni, A. Kannan, A. Criminisi and J. Winn In proc. CVPR Anchorage,
Saliency Aggregation: A Data- driven Approach Long Mai Yuzhen Niu Feng Liu Department of Computer Science, Portland State University Portland, OR,
SciFest Overview Neil Gannon. Outline Demonstrations using a Microsoft Kinect sensor – Image Manipulation Real-time Invisibility Background removal (green.
Final Presentation for EE7700 DVP Shenghua Wan and Kang zhang May, D View Simulation Based on Face Tracking.
CONTENT 1. Introduction to Kinect 2. Some Libraries for Kinect 3. Implement 4. Conclusion & Future works 1.
Tracking Systems in VR.
COMP24111: Machine Learning Ensemble Models Gavin Brown
11/05/15 How the Kinect Works Computational Photography Derek Hoiem, University of Illinois Photo frame-grabbed from:
RGB-D Images and Applications
3D Sensing 3D Shape from X Perspective Geometry Camera Model Camera Calibration General Stereo Triangulation 3D Reconstruction.
Product: Microsoft Kinect Team I Alex Styborski Brandon Sayre Brandon Rouhier Section 2B.
COM24111: Machine Learning Decision Trees Gavin Brown
Capstone Design Implementation of Depth Sensor Based on Structured Infrared Patterns June 11, 2013 School of Information and Communication Engineering,
Presenter: Jae Sung Park
1 2D TO 3D IMAGE AND VIDEO CONVERSION. INTRODUCTION The goal is to take already existing 2D content, and artificially produce the left and right views.
Maestro AI Vision and Design Overview Definitions Maestro: A naïve Sensorimotor Engine prototype. Sensorimotor Engine: Combining sensory and motor functions.
TOUCHLESS TOUCHSCREEN USER INTERFACE
Microsoft Kinect Jason Wong Pierce Nichols Rick Berggreen Tri Le.
Preliminary project assignment Smart house Natural User Interface for Business NUIT4B.
Creative Coding & the New Kinect
Southern Taiwan University Department of Electrical Engineering
A seminar on Touchless Touchscreen Technology
How Microsoft has leveraged
제 5 장 스테레오.
How the Kinect Works Computational Photography
Real-Time Human Pose Recognition in Parts from Single Depth Image
A New Approach to Track Multiple Vehicles With the Combination of Robust Detection and Two Classifiers Weidong Min , Mengdan Fan, Xiaoguang Guo, and Qing.
Data-driven Depth Inference from a Single Still Image
What is blue eyes ? aims on creating computational machines that have perceptual and sensory ability like those of human beings. interactive computer.
Multiple View Geometry for Robotics
Determining the Function Obtained from a Series of Transformations.
Presentation transcript:

Microsoft Kinect How does a machine infer body position?

Disclaimer A lot of the information presented comes from reverse engineering performed by a number of agencies. So they might not reflect true ground data, which is proprietary and thus not publicly available. 1

Microsoft Kinect  Series of motion sensing input devices for the Xbox  Launched in November 2010  Removes the need of a game controller  User interacts through gestures and spoken commands. 2

Two step Process  Compute Depth Map  Apply Machine Learning 3

Depth Map? 4

So how does it manage something so cool? 5

Projects a speckle pattern of Infrared Laser Light Depths Inferred from the distortions seen in the pattern 6

● Depth from Focus ○ What is far will be more blurry ○ Combine different focal lengths for X Axis and Y Axis ● Depth from Stereo ○ Combine image obtained from cameras in different positions ○ The speckle pattern projected from one point, observed from another (Cont) The cool combined with the Classic 7

We have the depth. What do we do next? Infer the Body Position, of course 8

Body Position Inference  Kinect starts with millions of labelled depth images  Learns a randomized decision forest, mapping depth images to body parts  This RDF takes days to be trained, but the process is made faster using parallel computing 9

Body Position Inference  Kinect uses the depth image and the RDF and computes a probability of a pixel belonging to each body part.  Uses mean shift algorithm to select the body parts and sketches a skeleton.  All of these computations are very fast. 10

Advantages  Works separately per frame as all the computations are very fast  Runs at 31 fps  Significantly more efficient and robust than body trackers  Works real-time with great results  Handles multiple players  Reported Accuracy: ~72% 11

References explained.html

Thank You Arhan Sibal Rudresh Pande