Hand Tracking for Virtual Object Manipulation

Slides:



Advertisements
Similar presentations
IMAGIS-GRAVIR / IMAG Augmented Reality Collaborative Environment: Calibration and Interactive Scene Editing Raphaël Grasset Xavier Decoret Jean-Dominique.
Advertisements

Computer vision: models, learning and inference Chapter 13 Image preprocessing and feature extraction.
Face Alignment with Part-Based Modeling
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
VisHap: Guangqi Ye, Jason J. Corso, Gregory D. Hager, Allison M. Okamura Presented By: Adelle C. Knight Augmented Reality Combining Haptics and Vision.
Object Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition l Panoramas,
Hybrid Position-Based Visual Servoing
Parallel Tracking and Mapping for Small AR Workspaces Vision Seminar
Vision Based Control Motion Matt Baker Kevin VanDyke.
Move With Me S.W Graduation Project An Najah National University Engineering Faculty Computer Engineering Department Supervisor : Dr. Raed Al-Qadi Ghada.
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Broadcast Court-Net Sports Video Analysis Using Fast 3-D Camera Modeling Jungong Han Dirk Farin Peter H. N. IEEE CSVT 2008.
Boundary matting for view synthesis Samuel W. Hasinoff Sing Bing Kang Richard Szeliski Computer Vision and Image Understanding 103 (2006) 22–32.
Motion Tracking. Image Processing and Computer Vision: 82 Introduction Finding how objects have moved in an image sequence Movement in space Movement.
Accurate Non-Iterative O( n ) Solution to the P n P Problem CVLab - Ecole Polytechnique Fédérale de Lausanne Francesc Moreno-Noguer Vincent Lepetit Pascal.
Multi video camera calibration and synchronization.
A Study of Approaches for Object Recognition
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Feature matching and tracking Class 5 Read Section 4.1 of course notes Read Shi and Tomasi’s paper on.
Vigilant Real-time storage and intelligent retrieval of visual surveillance data Dr Graeme A. Jones.
Feature tracking Class 5 Read Section 4.1 of course notes Read Shi and Tomasi’s paper on good features.
Pointing Based Object Localization CS223b Final Project Stanford University Bio-Robotics Lab Paul Nangeroni & Ashley Wellman March 17, 2008.
Presented by Pat Chan Pik Wah 28/04/2005 Qualifying Examination
1 Invariant Local Feature for Object Recognition Presented by Wyman 2/05/2006.
KLT tracker & triangulation Class 6 Read Shi and Tomasi’s paper on good features to track
Visual Screen: Transforming an Ordinary Screen into a Touch Screen Zhengyou Zhang & Ying Shan Vision Technology Group Microsoft Research
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
Computer Vision Group Computer Vision Projects (:,5) Vincenzo Caglioti Giacomo Boracchi, Simone Gasparini, Alessandro Giusti, Pierluigi Taddei.
Augmented Reality and 3D modelling Done by Stafford Joemat Supervised by Mr James Connan and Mr Mehrdad Ghaziasgar.
Knowledge Systems Lab JN 8/24/2015 A Method for Temporal Hand Gesture Recognition Joshua R. New Knowledge Systems Laboratory Jacksonville State University.
Kalman filter and SLAM problem
1 Intelligent Robotics Research Centre (IRRC) Department of Electrical and Computer Systems Engineering Monash University, Australia Visual Perception.
05 - Feature Detection Overview Feature Detection –Intensity Extrema –Blob Detection –Corner Detection Feature Descriptors Feature Matching Conclusion.
3D Fingertip and Palm Tracking in Depth Image Sequences
1 Interest Operators Harris Corner Detector: the first and most basic interest operator Kadir Entropy Detector and its use in object recognition SIFT interest.
Stereoscopic Imaging for Slow-Moving Autonomous Vehicle By: Alexander Norton Advisor: Dr. Huggins April 26, 2012 Senior Capstone Project Final Presentation.
Optical Tracking for VR Bertus Labuschagne Christopher Parker Russell Joffe.
Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.
A Local Adaptive Approach for Dense Stereo Matching in Architectural Scene Reconstruction C. Stentoumis 1, L. Grammatikopoulos 2, I. Kalisperakis 2, E.
Visual Perception PhD Program in Information Technologies Description: Obtention of 3D Information. Study of the problem of triangulation, camera calibration.
Topic 10 - Image Analysis DIGITAL IMAGE PROCESSING Course 3624 Department of Physics and Astronomy Professor Bob Warwick.
Imaging Geometry for the Pinhole Camera Outline: Motivation |The pinhole camera.
CVPR Workshop on RTV4HCI 7/2/2004, Washington D.C. Gesture Recognition Using 3D Appearance and Motion Features Guangqi Ye, Jason J. Corso, Gregory D. Hager.
CSCE 643 Computer Vision: Structure from Motion
3D Reconstruction Jeff Boody. Goals ● Reconstruct 3D models from a sequence of at least two images ● No prior knowledge of the camera or scene ● Use the.
Acquiring 3D models of objects via a robotic stereo head David Virasinghe Department of Computer Science University of Adelaide Supervisors: Mike Brooks.
Lecture 7: Features Part 2 CS4670/5670: Computer Vision Noah Snavely.
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
Pervasive Gaming with Mobile Devices KARNUNG LIANG SUPERVISOR: DR BRETT WILKINSON.
Real-Time Simultaneous Localization and Mapping with a Single Camera (Mono SLAM) Young Ki Baik Computer Vision Lab. Seoul National University.
Raquel A. Romano 1 Scientific Computing Seminar May 12, 2004 Projective Geometry for Computer Vision Projective Geometry for Computer Vision Raquel A.
Chapter 5 Multi-Cue 3D Model- Based Object Tracking Geoffrey Taylor Lindsay Kleeman Intelligent Robotics Research Centre (IRRC) Department of Electrical.
Looking at people and Image-based Localisation Roberto Cipolla Department of Engineering Research team
Knowledge Systems Lab JN 1/15/2016 Facilitating User Interaction with Complex Systems via Hand Gesture Recognition MCIS Department Knowledge Systems Laboratory.
Visual Odometry David Nister, CVPR 2004
Augmented Reality and 3D modelling Done by Stafford Joemat Supervised by Mr James Connan.
3D Reconstruction Using Image Sequence
Person Following with a Mobile Robot Using Binocular Feature-Based Tracking Zhichao Chen and Stanley T. Birchfield Dept. of Electrical and Computer Engineering.
Stereoscopic Imaging for Slow-Moving Autonomous Vehicle By: Alex Norton Advisor: Dr. Huggins February 28, 2012 Senior Project Progress Report Bradley University.
Instructor: Mircea Nicolescu Lecture 10 CS 485 / 685 Computer Vision.
Vision Based hand tracking for Interaction The 7th International Conference on Applications and Principles of Information Science (APIS2008) Dept. of Visual.
Augmented Reality and 3D modelling Done by Stafford Joemat Supervised by Mr James Connan and Mehrdad Ghaziasgar.
11/25/03 3D Model Acquisition by Tracking 2D Wireframes Presenter: Jing Han Shiau M. Brown, T. Drummond and R. Cipolla Department of Engineering University.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry
José Manuel Iñesta José Martínez Sotoca Mateo Buendía
Vehicle Segmentation and Tracking in the Presence of Occlusions
CAP 5415 Computer Vision Fall 2012 Dr. Mubarak Shah Lecture-5
Higher School of Economics , Moscow, 2016
An Introduction of Marker and Markerless In AR
Higher School of Economics , Moscow, 2016
Presentation transcript:

Hand Tracking for Virtual Object Manipulation Thibaut Weise Hi, I am Thibaut Weise. I’d like to present you my final year project which is about hand tracking for virtual object manipulation. Project Supervisor: Second Marker: Professor Guang-Zhong Yang Dr Duncan Gillies

Motivation Virtual and Augmented Environments Robotic Control Surgery Space, Deep Sea Why would we want to track hands? I’d like to give two examples where tracking hands would be beneficial: Virtual and Augmented Environments: Instead of using a mouse or wand, using the hands directly would be much more intuitive as it is the same movement as the user has to do for real world objects. It would make interaction for users much easier. Robot Control: Anywhere where hands are necessary for some work, but it is difficult to access, robotic arms can be used instead e.g. for surgery, mining, space, deep sea, etc. It would be beneficial to be able to control the robotic gripper directly by hand.

State of the Art EM-Tracking Acoustic Tracking Data Glove Optical

Optical Approaches Appearance-based Approach Model-based Approach Reconstruct 3D Hand Model Two methods: Minimize Model against Features Reconstruct from Features

Requirements Focus: Grab Gesture  Thumb and Index Finger Real-time Responsive Accurate

System Overview Calibration Marker-based Hand Tracking Marker-less Virtual Environment

Reconstruction Triangulation Requires Mapping between Pixels and 3D World Coordinates

Calibration Optimize known World Coordinates against detected Image Coordinates Planar Approach => Chessboard OpenCV Implementation Stereo Optimization

Calibration Accuracy Corner Detector Accuracy Why sub-pixel accurate?

X-Corner Detector Solution: Improved X-Corner Detector Saddle Point Product of Eigenvalues minimised: λ1λ2= fxxfyy-fxy2 Gradient zero: 0 = fx + s fxx + t fxy 0 = fy + s fxy + t fyy Recalculate using weighted derivatives of neighbourhood

Corner Detector Comparison Solution: Improved X-Corner Detector

Calibration Automation Improved Chessboard Detector

Calibration Tool

Hand Tracking Marker-based Marker-less

Marker Tracking 3 Markers Thumb, Index Finger and Hand Uses Improved X-Corner Detector Fast Accurate Simple Uses improved X-corner detector as it is fast, accurate and simple

Hand Reconstruction Find all potential Markers in Images Match Points using Epipolar Constraints Uses improved X-corner detector as it is fast, accurate and simple

Hand Reconstruction Reconstruct all possible hand configurations Different markers Validate hand constraints Uses improved X-corner detector as it is fast, accurate and simple

Hand Reconstruction Use best match based on temporal information Minimise Euclidean distance to predicted position Uses improved X-corner detector as it is fast, accurate and simple

Hand Reconstruction Initialisation based on Orientation Heuristic Uses improved X-corner detector as it is fast, accurate and simple

Marker-less Tracking Hand Localisation 2 Fingertips Reconstruct Hand Position Hand Localisation Skin Colour Detection We have no markers, so we need to reconstruct the hand differently. The hand position cannot found by direct feature point, as there is no feature! Reconstruct hand position from finger axes. First step is to localise the hand in the image. Skin colour is in a distinct region of a certain colour space. We use skin colour detection using a simple heuristic Rule.

Hand Localisation Background Subtraction - Skin + Background +

Hand Localisation Find Hand Contour We have no markers, so we need to reconstruct the hand differently. The markers were at the finger tips, so detecting the fingertips is effective. The hand position cannot found by direct feature point, as there is no feature! Reconstruct hand position from finger axes. First step is to localise the hand in the image. Skin colour is in a distinct region of a certain colour space. We use skin colour detection using a simple heuristic Rule.

Hand Reconstruction Find Peaks Calculate k-curvature: [C(i)-C(i-k)]T[C(i+k)-C(i)] Peaks where angle > 0o and < 90o Once we have a contour as a list of pixel positions we find the fingertips as peaks in the contour. For that we calculate the k-curvature at each contour pixel. The k-curvature is just the angle between the vector from the contour pixel k steps before and the vector to the contour pixel k steps after. If the angle is bigger than 90 and less than 180 degrees, that is the dot product is less than 0, then it is a potential peak.

Hand Reconstruction Match peaks using Epipolar Constraints  find possible Finger Tips Calculate Finger Axis

Hand Reconstruction Reconstruct hand position Validate hand constraints Use best hand based on temporal information

Evaluation Marker-based Marker-less

Virtual Environment Virtual Tower of Hanoi Virtual Hand => Grabber Fixed Dimensions Grab and Push Objects Perception Colour Coding Shadow

Implementation Application Video / Tracking Thread Thread Settings Video / Tracking Thread Draw / Settings Hand Update Virtual Environment

Demo

Conclusion Accurate semi-automatic Calibration Marker-based Tracking Marker-less Tracking Virtual Environment

Future Work Marker-based Tracking Marker-less Tracking Handle Occlusion Marker-less Tracking Refine Skin Colour Detection Use Gradient

Questions ? Who wants to try?