Consistent Visual Information Processing Axel Pinz EMT – Institute of Electrical Measurement and Measurement Signal Processing TU Graz – Graz University.

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
EU funded FP7: Oct 11 – Sep 14 Co-evolution of Future AR Mobile Platforms Paul Chippendale, Bruno Kessler Foundation FBK, Italy.
Feature Selection as Relevant Information Encoding Naftali Tishby School of Computer Science and Engineering The Hebrew University, Jerusalem, Israel NIPS.
CSCE643: Computer Vision Bayesian Tracking & Particle Filtering Jinxiang Chai Some slides from Stephen Roth.
Perceptive Context for Pervasive Computing Trevor Darrell Vision Interface Group MIT AI Lab.
For Internal Use Only. © CT T IN EM. All rights reserved. 3D Reconstruction Using Aerial Images A Dense Structure from Motion pipeline Ramakrishna Vedantam.
Institut für Elektrische Meßtechnik und Meßsignalverarbeitung Professor Horst Cerjak, Augmented Reality VU 2 Calibration Axel Pinz.
What is Virtual Reality? “A high-end user interface that involves real-time simulation and interaction through multiple sensorial channels.” (vision, sound,
Introduction To Tracking
Patch to the Future: Unsupervised Visual Prediction
Hybrid Position-Based Visual Servoing
Parallel Tracking and Mapping for Small AR Workspaces Vision Seminar
Vision Based Control Motion Matt Baker Kevin VanDyke.
Gerald Schweighofer RIGOROSUM Online SaM for GCM Institute of Electrical Measurement and Measurement Signal Processing Online Structure and.
Real-Time Tracking Axel Pinz Image Based Measurement Group EMT – Institute of Electrical Measurement and Measurement Signal Processing TU Graz – Graz University.
Daniel Shepard and Todd Humphreys
Asa MacWilliams Lehrstuhl für Angewandte Softwaretechnik Institut für Informatik Technische Universität München Dec Software.
Dana Cobzas-PhD thesis Image-Based Models with Applications in Robot Navigation Dana Cobzas Supervisor: Hong Zhang.
ICG Professor Horst Cerjak, Thomas Pock Variational Methods for 3D Reconstruction Thomas Pock 1, Chrisopher Zach 2 and Horst Bischof 1 1 Institute.
CH24 in Robotics Handbook Presented by Wen Li Ph.D. student Texas A&M University.
Tracking Objects with Dynamics Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 04/21/15 some slides from Amin Sadeghi, Lana Lazebnik,
Adam Rachmielowski 615 Project: Real-time monocular vision-based SLAM.
Tracking and Data Fusion for 4D Visualization Ulrich Neumann Suya You Computer Science Department Integrated Media Systems Center University of Southern.
A Study of Approaches for Object Recognition
Visual Odometry for Ground Vehicle Applications David Nister, Oleg Naroditsky, James Bergen Sarnoff Corporation, CN5300 Princeton, NJ CPSC 643, Presentation.
Theoretical Foundations of Multimedia Chapter 3 Virtual Reality Devices Non interactive Slow image update rate Simple image Nonengaging content and presentation.
Robotics for Intelligent Environments
Non-invasive Techniques for Human Fatigue Monitoring Qiang Ji Dept. of Electrical, Computer, and Systems Engineering Rensselaer Polytechnic Institute
InerVis Mobile Robotics Laboratory Institute of Systems and Robotics ISR – Coimbra Contact Person: Jorge Lobo Human inertial sensor:
Face Recognition and Retrieval in Video Basic concept of Face Recog. & retrieval And their basic methods. C.S.E. Kwon Min Hyuk.
Overview and Mathematics Bjoern Griesbach
Vision Guided Robotics
Jason Li Jeremy Fowers Ground Target Following for Unmanned Aerial Vehicles.
MACHINE VISION GROUP Multimodal sensing-based camera applications Miguel Bordallo 1, Jari Hannuksela 1, Olli Silvén 1 and Markku Vehviläinen 2 1 University.
Research Area B Leif Kobbelt. Communication System Interface Research Area B 2.
1 Intelligent Robotics Research Centre (IRRC) Department of Electrical and Computer Systems Engineering Monash University, Australia Visual Perception.
Driver’s View and Vehicle Surround Estimation using Omnidirectional Video Stream Abstract Our research is focused on the development of novel machine vision.
Real-Time Human Posture Reconstruction in Wireless Smart Camera Networks Chen Wu, Hamid Aghajan Wireless Sensor Network Lab, Stanford University, USA IPSN.
오 세 영, 이 진 수 전자전기공학과, 뇌연구센터 포항공과대학교
1. Introduction Motion Segmentation The Affine Motion Model Contour Extraction & Shape Estimation Recursive Shape Estimation & Motion Estimation Occlusion.
Characterizing activity in video shots based on salient points Nicolas Moënne-Loccoz Viper group Computer vision & multimedia laboratory University of.
A General Framework for Tracking Multiple People from a Moving Camera
Dynamic 3D Scene Analysis from a Moving Vehicle Young Ki Baik (CV Lab.) (Wed)
MURI: Integrated Fusion, Performance Prediction, and Sensor Management for Automatic Target Exploitation 1 Dynamic Sensor Resource Management for ATE MURI.
Features-based Object Recognition P. Moreels, P. Perona California Institute of Technology.
Vision-based human motion analysis: An overview Computer Vision and Image Understanding(2007)
Chapter 5 Multi-Cue 3D Model- Based Object Tracking Geoffrey Taylor Lindsay Kleeman Intelligent Robotics Research Centre (IRRC) Department of Electrical.
HCI 입문 Graphics Korea University HCI System 2005 년 2 학기 김 창 헌.
Ubitrack - Ubiquitous Tracking for Augmented Reality Joe Newman For more information
Augmented Reality Authorized By: Miss.Trupti Pardeshi. NDMVP, Comp Dept. Augmented Reality 1/ 23.
Looking at people and Image-based Localisation Roberto Cipolla Department of Engineering Research team
CHAPTER 1: Introduction. 2 Why “Learn”? Machine learning is programming computers to optimize a performance criterion using example data or past experience.
It Starts with iGaze: Visual Attention Driven Networking with Smart Glasses It Starts with iGaze: Visual Attention Driven Networking with Smart Glasses.
IEEE International Conference on Multimedia and Expo.
Visual Odometry for Ground Vehicle Applications David Nistér, Oleg Naroditsky, and James Bergen Sarnoff Corporation CN5300 Princeton, New Jersey
1 Review and Summary We have covered a LOT of material, spending more time and more detail on 2D image segmentation and analysis, but hopefully giving.
University of Pennsylvania 1 GRASP Control of Multiple Autonomous Robot Systems Vijay Kumar Camillo Taylor Aveek Das Guilherme Pereira John Spletzer GRASP.
CIRP Annals - Manufacturing Technology 60 (2011) 1–4 Augmented assembly technologies based on 3D bare-hand interaction S.K. Ong (2)*, Z.B. Wang Mechanical.
A. M. R. R. Bandara & L. Ranathunga
Paper – Stephen Se, David Lowe, Jim Little
Tracking Objects with Dynamics
Florian Shkurti, Ioannis Rekleitis, Milena Scaccia and Gregory Dudek
Video-based human motion recognition using 3D mocap data
Machine Learning Feature Creation and Selection
Combining Geometric- and View-Based Approaches for Articulated Pose Estimation David Demirdjian MIT Computer Science and Artificial Intelligence Laboratory.
George Bebis and Wenjing Li Computer Vision Laboratory
Sensor Fusion Localization and Navigation for Visually Impaired People
Introduction to Object Tracking
Tracking Many slides adapted from Kristen Grauman, Deva Ramanan.
Presentation transcript:

Consistent Visual Information Processing Axel Pinz EMT – Institute of Electrical Measurement and Measurement Signal Processing TU Graz – Graz University of Technology

“Consistency” Active vision systems / 4D data streams Imprecision Ambiguity Contradiction Multiple visual information

This Talk: Consistency in Active vision systems: –Active fusion –Active object recognition Immersive 3D HCI: –Augmented reality –Tracking in VR/AR

AR as Testbed Consistent perception in 4D: Space –Registration –Tracking Time –Lag-free –Prediction

Agenda Active fusion Consistency Applications –Active object recognition –Tracking in VR/AR Conclusions

Active Fusion Simple top level decision-action-fusion loop:

Active Fusion (2) Fusion schemes –Probabilistic –Possibilistic (fuzzy) –Evidence theoretic (Dempster & Shafer)

Probabilistic Active Fusion N measurements, sensor inputs: m i M hypotheses: o j, O = {o 1, …, o M } Bayes formula: Use entropy H(O) to measure the quality of P(O)

Probabilistic Active Fusion (2) Flat distribution: P(o j )=const.  H max Measurements can be: difficult, expensive, N can be prohibitively large, …  Find iterative strategy to minimize H(O) Pronounced distribution: P(o c ) = 1; P(o j ) = 0, j  c  H = 0

Probabilistic Active Fusion (3) Start with A  1 measurements: P(o j |m 1, …,m A ), H A Iteratively take more measurements: m A+1, …,m B Until: P(o j |m 1, …,m B ), H B  Threshold

Summary: Active Fusion Multiple (visual) information, many sensors, measurements,… Selection of information sources Maximize information content / quality Optimize effort (number / cost of measurements, …) Information gain by entropy reduction

Summary: Active Fusion (2) Active systems (robots, mobile cameras) –Sensor planning –Control –Interaction with the scene “Passive” systems (video, wearable,…) –Filtering –Selection of sensors / measurements

Consistency Consistency vs. Ambiguity –Unimodal subsets O k Representations –Distance measures

Consistent Subsets Hypotheses O = {o 1,…, o M } Ambiguity: P(O) is multimodal Consistent unimodal subsets O k  O Application domains Support of hypotheses Outlier rejection Benefits:

Distance Measures Depend on representations, e.g.: Pixel-levelSSD, correlation, rank EigenspaceEuclidean 3D modelsEuclidean Feature-basedMahalanobis, … SymbolicMutual information GraphsSubgraph isomorphism

Mutual Information Shannon´s measure of mutual information: O = {o 1,…, o M } A  O, B  O I(A,B) = H(A) + H(B) – H(A,B)

Applications Active object recognition –Videos –Details Tracking in VR / AR –Landmark definition / acquisition –Real-time tracking

Active vision laboratory

Active Object Recognition

Active Object Recognition in Parametric Eigenspace Classifier for a single view Pose estimation per view Fusion formalism View planning formalism Estimation of object appearance at unexplored viewing positions

Applications  Active object recognition –Videos –Details  Control of active vision systems Tracking in VR / AR –Landmark definition / acquisition –Real-time tracking  Selection, combination, evaluation  Constraining of huge spaces

Landmark Definition / Acquisition cornersblobsnatural landmarks What is a “landmark” ?

Automatic Landmark Acquisition Capture a dataset of the scene: –calibrated stereo rig –trajectory (by magnetic tracking) –n stereo pairs Process this dataset –visually salient landmarks for tracking

Automatic Landmark Acquisition visually salient landmarks for tracking salient points in 2D image 3D reconstruction clusters in 3D: –compact, many points –consistent feature descriptions cluster centers  landmarks

Processing Scheme

Office Scene

Office Scene - Reconstruction

Unknown Scene Real-Time Tracking Landmark Acquisition

Real-Time Tracking Measure position and orientation of object(s) Obtain trajectories of object(s) Stationary observer – “outside-in” –Vision-based Moving observer, egomotion – “inside-out” –Hybrid Degrees of Freedom – DoF –3 DoF (mobile robot) –6 DoF (head and device tracking in AR)

Outside-in Tracking (1) stereo-rig IR-illumination wireless 1 marker/device: 3 DoF 2 markers: 5 DoF 3 markers: 6 DoF devices

Outside-in Tracking (2)

Consistent Tracking (1) Complexity –Many targets –Exhaustive search vs. Real-time Occlusion –Redundancy (targets | cameras) Ambiguity in 3D –Constraints

Consistent Tracking (2) Dynamic interpretation tree –Geometric / spatial consistency Local constraints –Multiple interpretations can happen –Global consistency is impossible Temporal consistency –Filtering, prediction

Consistent Tracking (3)

Hybrid Inside-Out Tracking (1) 3 accelerometers 3 gyroscopes signal processing interface Inertial Tracker

Hybrid Inside-Out Tracking (2) complementary sensors fusion

Summary: Consistency in Active vision systems: –Active fusion –Active object recognition Immersive 3D HCI: –Augmented reality –Tracking in VR/AR

Conclusion Consistent processing of visual information can significantly improve the performance of active and real-time vision systems

Acknowledgement Thomas Auer, Hermann Borotschnig, Markus Brandner, Harald Ganster, Peter Lang, Lucas Paletta, Manfred Prantl, Miguel Ribo, David Sinclair Christian Doppler Gesellschaft, FFF, FWF, Kplus VRVis, EU TMR Virgo