Learning about Objects

Slides:



Advertisements
Similar presentations
Feature extraction: Corners
Advertisements

Road-Sign Detection and Recognition Based on Support Vector Machines Saturnino, Sergio et al. Yunjia Man ECG 782 Dr. Brendan.
Evaluating Color Descriptors for Object and Scene Recognition Koen E.A. van de Sande, Student Member, IEEE, Theo Gevers, Member, IEEE, and Cees G.M. Snoek,
Perception and Perspective in Robotics Paul Fitzpatrick MIT Computer Science and Artificial Intelligence Laboratory Humanoid Robotics Group Goal To build.
The DayOne project: how far can a robot develop in 24 hours? Paul Fitzpatrick MIT CSAIL.
Computer Vision Lecture 16: Region Representation
Vision Based Control Motion Matt Baker Kevin VanDyke.
Color Image Processing
Intelligent Systems Lab. Extrinsic Self Calibration of a Camera and a 3D Laser Range Finder from Natural Scenes Davide Scaramuzza, Ahad Harati, and Roland.
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
1 of 25 1 of 22 Blind-Spot Experiment Draw an image similar to that below on a piece of paper (the dot and cross are about 6 inches apart) Close your right.
Feature extraction: Corners 9300 Harris Corners Pkwy, Charlotte, NC.
1 Color Segmentation: Color Spaces and Illumination Mohan Sridharan University of Birmingham
A Study of Approaches for Object Recognition
CS 561, Sessions 27 1 Towards intelligent machines Thanks to CSCI561, we now know how to… - Search (and play games) - Build a knowledge base using FOL.
Highlights Lecture on the image part (10) Automatic Perception 16
Lecture 3a: Feature detection and matching CS6670: Computer Vision Noah Snavely.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Lecture 6: Feature matching and alignment CS4670: Computer Vision Noah Snavely.
Circular Augmented Rotational Trajectory (CART) Shape Recognition & Curvature Estimation Presentation for 3IA 2007 Russel Ahmed Apu & Dr. Marina Gavrilova.
Computer vision.
The Whole World in Your Hand: Active and Interactive Segmentation The Whole World in Your Hand: Active and Interactive Segmentation – Artur Arsenio, Paul.
1. Introduction Motion Segmentation The Affine Motion Model Contour Extraction & Shape Estimation Recursive Shape Estimation & Motion Estimation Occlusion.
OBJECT RECOGNITION. The next step in Robot Vision is the Object Recognition. This problem is accomplished using the extracted feature information. The.
CSCE 5013 Computer Vision Fall 2011 Prof. John Gauch
Lecture 4: Feature matching CS4670 / 5670: Computer Vision Noah Snavely.
Vision & Recognition. From a different direction At different times, particularly if it has been modified in the interval In different light A particular.
Computer Science Department Pacific University Artificial Intelligence -- Computer Vision.
Acquiring 3D models of objects via a robotic stereo head David Virasinghe Department of Computer Science University of Adelaide Supervisors: Mike Brooks.
Natural Tasking of Robots Based on Human Interaction Cues Brian Scassellati, Bryan Adams, Aaron Edsinger, Matthew Marjanovic MIT Artificial Intelligence.
Feature extraction: Corners 9300 Harris Corners Pkwy, Charlotte, NC.
1 Artificial Intelligence: Vision Stages of analysis Low level vision Surfaces and distance Object Matching.
1 Research Question  Can a vision-based mobile robot  with limited computation and memory,  and rapidly varying camera positions,  operate autonomously.
Robotic Chapter 8. Artificial IntelligenceChapter 72 Robotic 1) Robotics is the intelligent connection of perception action. 2) A robotic is anything.
Object Lesson: Discovering and Learning to Recognize Objects Object Lesson: Discovering and Learning to Recognize Objects – Paul Fitzpatrick – MIT CSAIL.
Corner Detection & Color Segmentation CSE350/ Sep 03.
Exploiting cross-modal rhythm for robot perception of objects Artur M. Arsenio Paul Fitzpatrick MIT Computer Science and Artificial Intelligence Laboratory.
Features, Feature descriptors, Matching Jana Kosecka George Mason University.
APECE-505 Intelligent System Engineering Basics of Digital Image Processing! Md. Atiqur Rahman Ahad Reference books: – Digital Image Processing, Gonzalez.
Reference books: – Digital Image Processing, Gonzalez & Woods. - Digital Image Processing, M. Joshi - Computer Vision – a modern approach, Forsyth & Ponce.
Colour and Texture. Extract 3-D information Using Vision Extract 3-D information for performing certain tasks such as manipulation, navigation, and recognition.
Giorgio Metta · Paul Fitzpatrick Humanoid Robotics Group MIT AI Lab Better Vision through Manipulation Manipulation Manipulation.
Robotics Chapter 6 – Machine Vision Dr. Amit Goradia.
Image features and properties. Image content representation The simplest representation of an image pattern is to list image pixels, one after the other.
Feel the beat: using cross-modal rhythm to integrate perception of objects, others, and self Paul Fitzpatrick and Artur M. Arsenio CSAIL, MIT.
AN ACTIVE VISION APPROACH TO OBJECT SEGMENTATION – Paul Fitzpatrick – MIT CSAIL.
Manipulation in Human Environments
Another Example: Circle Detection
A Plane-Based Approach to Mondrian Stereo Matching
Processing visual information for Computer Vision
San Diego May 22, 2013 Giovanni Saponaro Giampiero Salvi
Color Image Processing
Color Image Processing
Modeling the development of mirror neurons Problem Solution
José Manuel Iñesta José Martínez Sotoca Mateo Buendía
Nearest-neighbor matching to feature database
Color Image Processing
Manipulation in Human Environments
Head pose estimation without manual initialization
New horizons in the artificial vision
Opportunistic Perception
Vehicle Segmentation and Tracking in the Presence of Occlusions
Perception and Perspective in Robotics
Nearest-neighbor matching to feature database
Paul Fitzpatrick Humanoid Robotics Group MIT AI Lab
Color Image Processing
Better Vision through Experimental Manipulation
Aim of the project Take your image Submit it to the search engine
Object Recognition Today we will move on to… April 12, 2018
Color Image Processing
Presentation transcript:

Learning about Objects by Experiment – Paul Fitzpatrick – MIT Artificial Intelligence Laboratory

    machine perception algorithms robust perception training data parameter tweaking  

self-training perceptual system algorithms opportunities training data   robust perception more algorithms  

making opportunities In robotics, vision is often used to guide manipulation But manipulation can also guide vision Important for… Experimentation progressing when perception is ambiguous Correction recovering when perception is misleading Development bootstrapping when perception is primitive

a simple scene? Edges of table and cube overlap Color of cube and table are poorly separated Cube has misleading surface pattern Maybe some cruel grad-student faked the cube with paper, or glued it to the table

active segmentation Object boundaries are not always easy to detect visually

active segmentation Object boundaries are not always easy to detect visually Solution: Cog sweeps through ambiguous area

active segmentation Object boundaries are not always easy to detect visually Solution: Cog sweeps through ambiguous area Resulting object motion helps segmentation

active segmentation Object boundaries are not always easy to detect visually Solution: Cog sweeps through ambiguous area Resulting object motion helps segmentation Robot can learn to recognize and segment object without further contact

active segmentation

active segmentation

integrated behavior

segmentation examples

segmentation examples table car

boundary fidelity cube car bottle ball 0.05 0.1 0.15 0.2 0.25 0.3 0.04 0.08 0.12 0.16 cube car bottle ball (square root of) second Hu moment measure of anisiotropy measure of size (area at poking distance)

affordance exploitation poking object segmentation edge catalog object detection (recognition, localization, contact-free segmentation) affordance exploitation (rolling) manipulator detection (robot, human)

poking object segmentation

opportunities for segmentation robot, poking (Paul Fitzpatrick, Giorgio Metta) wearable, illumination (Charlie Kemp) camera, periodicity (Artur Arsenio)

segmentation on a wearable System detects when wearer reaches for an object, requests wearer to hold it up, then illuminates it

segmentation by watching a human System detects periodic motion – waving, tapping, etc. – and extracts seed points for segmentation

poking object segmentation edge catalog

sampling oriented regions

sample samples

most frequent samples 1st 21st 41st 61st 81st 101st 121st 141st 161st

Red = horizontal Green = vertical some tests Red = horizontal Green = vertical

natural images 00 22.50 900 22.50 450 22.50 450 22.50

poking object segmentation edge catalog object detection (recognition, localization, contact-free segmentation)

geometry+appearance Advantages: more selective; fast angles + colors + relative sizes invariant to scale, translation, In-plane rotation Advantages: more selective; fast Disadvantages: edges can be occluded; 2D method Property: no need for offline training

in operation

yellow on yellow

open object recognition robot’s current view recognized object (as seen during poking) pokes, segments ball sees ball, “thinks” it is cube correctly differentiates ball and cube

open object recognition

manipulator detection poking object segmentation edge catalog manipulator detection (robot, human) object detection (recognition, localization, contact-free segmentation)

manipulator recognition

affordance exploitation object segmentation poking affordance exploitation (rolling) object segmentation edge catalog manipulator detection (robot, human) object detection (recognition, localization, contact-free segmentation)

objects roll in different ways a bottle it rolls along its side a toy car it rolls forward a toy cube it doesn’t roll easily a ball it rolls in any direction

affordance exploitation

mimicry test Demonstration by human Mimicry in similar situation Mimicry when object is rotated Invoking the object’s natural rolling affordance Going against the object’s natural rolling affordance

mimicry test

affordance exploitation object segmentation poking affordance exploitation (rolling) object segmentation edge catalog manipulator detection (robot, human) object detection (recognition, localization, contact-free segmentation)

referring to objects

beyond objects familiar activities use constraint of familiar activity to discover unfamiliar entity used within it reveal the structure of unfamiliar activities by tracking familiar entities into and through them familiar entities (objects, actors, properties, …)

learning activity structure /ylo/ /lft/ /grin/ /ryt/ /ylo/ /lft/ /grin/ /ryt/ /ylo/ /lft/ /grin/ /ryt/ /ylo/ /lft/ /grin/ /ryt/ /ylo/ /lft/ /grin/ /ryt/ /ylo/ /lft/ /grin/ /ryt/ /ylo/ /lft/ /grin/ /ryt/ /ylo/ /lft/ /grin/ /ylo/ /lft/ /ylo/ /lft/ /ryt/ Extending Vocabulary /lft/ /yn/ /ryt/ Extending Vocabulary /lft/ /ylo/ /ryt/ /grin/ /lft/ /ylo/ /ryt/

conclusions active segmentation (through contact) appearance catalog (for oriented features) open object recognition for correction, enrollment affordance recognition (for rolling) open speech recognition (for isolated words) virtuous circle of development learning about and through activity