Humanoid Robot In Our World Toward humanoid manipulation in human-centred environments Presented By Yu Gu (yg2466)

Slides:



Advertisements
Similar presentations
Presented by Xinyu Chang
Advertisements

Evaluating Color Descriptors for Object and Scene Recognition Koen E.A. van de Sande, Student Member, IEEE, Theo Gevers, Member, IEEE, and Cees G.M. Snoek,
RGB-D object recognition and localization with clutter and occlusions Federico Tombari, Samuele Salti, Luigi Di Stefano Computer Vision Lab – University.
Human Identity Recognition in Aerial Images Omar Oreifej Ramin Mehran Mubarak Shah CVPR 2010, June Computer Vision Lab of UCF.
System Integration and Experimental Results Intelligent Robotics Research Centre (IRRC) Department of Electrical and Computer Systems Engineering Monash.
Object Recognition & Model Based Tracking © Danica Kragic Tracking system.
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
A Versatile Depalletizer of Boxes Based on Range Imagery Dimitrios Katsoulas*, Lothar Bergen*, Lambis Tassakos** *University of Freiburg **Inos Automation-software.
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Exchanging Faces in Images SIGGRAPH ’04 Blanz V., Scherbaum K., Vetter T., Seidel HP. Speaker: Alvin Date: 21 July 2004.
ECE 7340: Building Intelligent Robots QUALITATIVE NAVIGATION FOR MOBILE ROBOTS Tod S. Levitt Daryl T. Lawton Presented by: Aniket Samant.
A Study of Approaches for Object Recognition
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Feature matching and tracking Class 5 Read Section 4.1 of course notes Read Shi and Tomasi’s paper on.
Face Recognition Based on 3D Shape Estimation
3D Hand Pose Estimation by Finding Appearance-Based Matches in a Large Database of Training Views
Scale Invariant Feature Transform (SIFT)
The Terrapins Computer Vision Laboratory University of Maryland.
Overview Introduction to local features
Distinctive Image Features from Scale-Invariant Keypoints By David G. Lowe, University of British Columbia Presented by: Tim Havinga, Joël van Neerbos.
1 Intelligent Robotics Research Centre (IRRC) Department of Electrical and Computer Systems Engineering Monash University, Australia Visual Perception.
Perception Introduction Pattern Recognition Image Formation
International Conference on Computer Vision and Graphics, ICCVG ‘2002 Algorithm for Fusion of 3D Scene by Subgraph Isomorphism with Procrustes Analysis.
Fitting: The Hough transform. Voting schemes Let each feature vote for all the models that are compatible with it Hopefully the noise features will not.
December 4, 2014Computer Vision Lecture 22: Depth 1 Stereo Vision Comparing the similar triangles PMC l and p l LC l, we get: Similarly, for PNC r and.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
A Region Based Stereo Matching Algorithm Using Cooperative Optimization Zeng-Fu Wang, Zhi-Gang Zheng University of Science and Technology of China Computer.
Fitting: The Hough transform
1 Research Question  Can a vision-based mobile robot  with limited computation and memory,  and rapidly varying camera positions,  operate autonomously.
Action and Gait Recognition From Recovered 3-D Human Joints IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS— PART B: CYBERNETICS, VOL. 40, NO. 4, AUGUST.
Source: Computer Vision and Pattern Recognition Workshops (CVPRW), 2010 IEEE Computer Society Conference on Author: Paucher, R.; Turk, M.; Adviser: Chia-Nian.
Chapter 5 Multi-Cue 3D Model- Based Object Tracking Geoffrey Taylor Lindsay Kleeman Intelligent Robotics Research Centre (IRRC) Department of Electrical.
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
Vision and SLAM Ingeniería de Sistemas Integrados Departamento de Tecnología Electrónica Universidad de Málaga (Spain) Acción Integrada –’Visual-based.
Jack Pinches INFO410 & INFO350 S INFORMATION SCIENCE Computer Vision I.
CSE 185 Introduction to Computer Vision Feature Matching.
Chapter 8. Learning of Gestures by Imitation in a Humanoid Robot in Imitation and Social Learning in Robots, Calinon and Billard. Course: Robots Learning.
Visual Odometry David Nister, CVPR 2004
Stereo Vision Local Map Alignment for Robot Environment Mapping Computer Vision Center Dept. Ciències de la Computació UAB Ricardo Toledo Morales (CVC)
Mobile Robot Localization and Mapping Using Range Sensor Data Dr. Joel Burdick, Dr. Stergios Roumeliotis, Samuel Pfister, Kristo Kriechbaum.
Hough Transform CS 691 E Spring Outline Hough transform Homography Reading: FP Chapter 15.1 (text) Some slides from Lazebnik.
Toward humanoid manipulation in human-centered environments T. Asfour, P. Azad, N. Vahrenkamp, K. Regenstein, A. Bierbaum, K. Welke, J. Schroder, R. Dillmann.
CSCI 631 – Foundations of Computer Vision March 15, 2016 Ashwini Imran Image Stitching Link: singhashwini.mesinghashwini.me.
Vision-Guided Humanoid Footstep Planning for Dynamic Environments
Processing visual information for Computer Vision
Visual homing using PCA-SIFT
A. M. R. R. Bandara & L. Ranathunga
Introduction to Skin and Face Detection
First-person Teleoperation of Humanoid Robots
Andreas Hermann, Felix Mauch, Sebastian Klemm, Arne Roennau
Paper – Stephen Se, David Lowe, Jim Little
José Manuel Iñesta José Martínez Sotoca Mateo Buendía
Recognizing Deformable Shapes
GESTURE CONTROLLED ROBOTIC ARM
Manipulation in Human Environments
Real-Time Human Pose Recognition in Parts from Single Depth Image
Factors that Influence the Geometric Detection Pattern of Vehicle-based Licence Plate Recognition Systems Martin Rademeyer Thinus Booysen, Arno Barnard.
Common Classification Tasks
Aim of the project Take your image Submit it to the search engine
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
Outline H. Murase, and S. K. Nayar, “Visual learning and recognition of 3-D objects from appearance,” International Journal of Computer Vision, vol. 14,
KFC: Keypoints, Features and Correspondences
Liyuan Li, Jerry Kah Eng Hoe, Xinguo Yu, Li Dong, and Xinqi Chu
Paper Reading Dalong Du April.08, 2011.
CSE 185 Introduction to Computer Vision
Multi-Information Based GCPs Selection Method
Humanoid Motion Planning for Dual-Arm Manipulation and Re-Grasping Tasks Nikolaus Vahrenkamp, Dmitry Berenson, Tamim Asfour, James Kuffner, Rudiger Dillmann.
Presented by Xu Miao April 20, 2005
Unsupervised Perceptual Rewards For Imitation Learning
Chapter 4 . Trajectory planning and Inverse kinematics
Presentation transcript:

Humanoid Robot In Our World Toward humanoid manipulation in human-centred environments Presented By Yu Gu (yg2466)

Outline Introduction & Motivation ARMAR III – The robot System Motion Planning Recognition and Localization Grasping

Introduction Current Research Area Human – Humanoid Interaction / Cooperation Human – centred Environment : Household Skill – Manipulative Perceptive Communicative

Motivation Put Humanoid Robots in our environment Let’s coexist!

Motion Planning The Robot Control Architecture Recognition Localization Grasp Analysis System

ARMAR-III : The Robot Design Consideration: Mimic Human Sensory & Sensory Motor capabilities Deal with Household Environments Versatile Sensory Motor: Sensorimotor skills involve the process of receiving sensory messages (sensory input) and producing a response (motor output). (mention during pre)

ARMAR-III : The Robot Design Specs: Upper Body Light & Modular Similar Size Similar Proportion Lower Body Holonomic movability

ARMAR-III : The Robot Components Head Eye Neck Arms Torso Hand Presentation ques: Head : 7 DOF + 2 eyes (3 DOF), Each eye = two digital color camera, 1 wide-angle lens + 1 narrow-angle lens Mounted on : 4 DOF neck Acoustic: head + microphone array (6) + inertial sensor Upper Body: 33 DOF, 14 DOF arm 3 DOF Torso Arm: shoulder 3 DOF, Elbow 2 DOF, Wrist 2 DOF – Five fingered hand (8 DOF) Platform: Wheel-based Platform

ARMAR-III : The Robot Components Platform Locomotion Omniwheels Sensor System Laser Scanner Optical Encoder Presentation ques: Head : 7 DOF + 2 eyes (3 DOF), Each eye = two digital color camera, 1 wide-angle lens + 1 narrow-angle lens Mounted on : 4 DOF neck Acoustic: head + microphone array (6) + inertial sensor Upper Body: 33 DOF, 14 DOF arm 3 DOF Torso Arm: shoulder 3 DOF, Elbow 2 DOF, Wrist 2 DOF – Five fingered hand (8 DOF) Platform: Wheel-based

ARMAR-III : The Robot

Motion Planning The Robot Control Architecture Recognition Localization Grasp Analysis System

ARMAR-III Control Architecture Task Planning Level Synchronization and Coordination Level Sensor-Motor Level

ARMAR-III Control Architecture Tasks need to be decomposed Subtasks are scheduled, executed and synchronized

ARMAR-III Control Architecture Task Planning Level Highest Level Responsible for: Task scheduling Resource\Skill Management Generate Subtasks Task Coordination Level Activates Sequential/Parallel Actions

ARMAR-III Control Architecture Task Execution Level Execute Commands Local Models: Active Model (short-term memory)

ARMAR-III Computer Architecture

Motion Planning The Robot Control Architecture Recognition Localization Grasp Analysis System

ARMAR-III Motion Planning Demo 1 2 Enlarged Robot Models Ensure Collision-free Paths Using Enlarged Robot Models Result Lazy Collision Checking

ARMAR-III Motion Planning Fast & Adaptive to Changing Environment Previous Methods: not for humanoid Bubbles ? - slow current: RRT Resolution Parameter Also… Ensure Collision-free Paths Show RRT video in case someone does not know what is RRT

ARMAR-III Motion Planning Enlarged Robot Models

ARMAR-III Motion Planning Planning Enlarged Robot Models Lower bond Two step Lazy Collision Checking Normal RRT Sampling Validation Validation using enlarged Detour if segment fails

ARMAR-III Motion Planning Result

Motion Planning The Robot Control Architecture Recognition Localization Grasp Analysis System

ARMAR-III Recognition & Localization Based on shape Recognition & localization Based on Texture Global Appearance-based object recognition system Region Processing 6D localization Local Appearance-based object recognition system Feature Extraction Object Recognition 2D & 6D localization P. Azad, T. Asfour, R. Dillmann, Combining appearance-based and model-based methods for real-time object recognition and 6Dlocalization, in: IEEE/RSJ International Conference on Intelligent Robots and Systems, Beijing, China, 2006. S. Nayar, S. Nene, and H. Murase, “Real-time 100 object recognition system,” in International Conference on Robotics and Automation (ICRA), vol. 3, Minneapolis, USA, 1996, pp. 2321–2325.

ARMAR-III Recognition & Localization Based on shape Segmentation : Color segmentation in HSV Limits: uses solid color objects Region Processing Pipeline

ARMAR-III Recognition & Localization Region Processing Pipeline S1: Normalization S2: Gradient Image Normalized Representation in size Robust; Less ambiguous; Good for Various light conditions P. Azad, T. Asfour, R. Dillmann, Combining appearance-based and model-based methods for real-time object recognition and 6Dlocalization, in: IEEE/RSJ International Conference on Intelligent Robots and Systems, Beijing, China, 2006. S. Nayar, S. Nene, and H. Murase, “Real-time 100 object recognition system,” in International Conference on Robotics and Automation (ICRA), vol. 3, Minneapolis, USA, 1996, pp. 2321–2325.

ARMAR-III Recognition & Localization 6D localization Orientation Position Calculate Position & Orientation independently Position: Triangulation Centroid Orientation: Retrieve View From DB Acceleration PCA 3D Model -> View P. Azad, T. Asfour, R. Dillmann, Combining appearance-based and model-based methods for real-time object recognition and 6Dlocalization, in: IEEE/RSJ International Conference on Intelligent Robots and Systems, Beijing, China, 2006. S. Nayar, S. Nene, and H. Murase, “Real-time 100 object recognition system,” in International Conference on Robotics and Automation (ICRA), vol. 3, Minneapolis, USA, 1996, pp. 2321–2325.

ARMAR-III Recognition & Localization Based on Texture Feature Calculation Shi-Tomasi Feature – View Set Maximally Stable Extremal Regions (MSER) combined with LAF SIFT Features - BEST

ARMAR-III Recognition & Localization SIFT descriptor Rotation Invariant Skew&Depth Invariant to some degree Rotational Angle & Feature Vector Gradient based SIFT Feature D.G. Lowe, Object recognition from local scale-invariant features, in:International Conference on Computer Vision, ICCV, Corfu, Greece, 1999, pp. 1150–1517.

ARMAR-III Recognition & Localization General Hough Transform find imperfect instances of objects within a certain class of shapes by a voting procedure. This voting procedure is carried out in a parameter space, from which object candidates are obtained as local maxima in a so-called accumulator space that is explicitly constructed by the algorithm for computing the Hough transform. Voting – Rotative Information Object Recognition 2D Localization D.G. Lowe, Object recognition from local scale-invariant features, in:International Conference on Computer Vision, ICCV, Corfu, Greece, 1999, pp. 1150–1517. https://en.wikipedia.org/wiki/Hough_transform

ARMAR-III Recognition & Localization 6D Localization POSIT Algorithm D.G. Lowe, Object recognition from local scale-invariant features, in:International Conference on Computer Vision, ICCV, Corfu, Greece, 1999, pp. 1150–1517. https://en.wikipedia.org/wiki/Hough_transform

ARMAR-III Recognition & Localization 6D Localization Limitation Correctness depends on 2D correspondence only Depth infor – sensitive to error in 2D coord Approach for cuboids Determine Highly textured points Determine Correspondence Determine Highly textured points Fit a 3D plane Calculate 3D points D. DeMenthon, L. Davis, D. Oberkampf, Iterative pose estimation using coplanar points, in: International Conference on Computer Vision and Pattern Recognition, CVPR, 1993, pp. 626–627.

ARMAR-III Recognition & Localization 6D Localization Approach for cuboids Determine Highly textured points Calculate 3D points Determine Correspondence Determine Highly textured points Fit a 3D plane D. DeMenthon, L. Davis, D. Oberkampf, Iterative pose estimation using coplanar points, in: International Conference on Computer Vision and Pattern Recognition, CVPR, 1993, pp. 626–627.

Motion Planning The Robot Control Architecture Recognition Localization Grasp Analysis System

ARMAR-III Grasping Demo 1 2 3 4 5 Central Idea: DB for 3D models Components: Global Model database Offline Grasp Analyzer Online Visual Procedure to identify objects in stereo images Select Grasp – Simulation Using GraspIt!

ARMAR-III Grasping Components: Global Model database – Model + Grasp Offline Grasp Analyzer – Computer Stable Grasp Online Visual Procedure to identify objects in stereo images Match Model (Recognition) & Find loc + pose Select Grasp

ARMAR-III Grasping Offline Grasp Analysis Limitation: Inaccuracy & Uncertainty of the object: Location Composition of a Grasp Grasp Type Grasp Starting Point (GSP) Approaching Direction Hand Orientation

ARMAR-III Grasping Grasp Centre Point (GCP) Virtual Point Align with GSP Practical Application Store Grasp in order

Thoughts On Paper Pros: Cons: Provided integrated humanoid platform and system for further development Not only for Research but also for real applications Has Vision System + Path Planner + Offline Grasp Analyser Cons: For recognition: Segmentation is not robust across diverse color range Pre-Stored Model + Grasp

Thoughts On Paper What can be done ? Train of Thought On-board Powerful compute engine Use Deep learning to adapt to novel objects Train of Thought Why not use assumption ?

Thank You For Listening!