Object Lesson: Discovering and Learning to Recognize Objects Object Lesson: Discovering and Learning to Recognize Objects – Paul Fitzpatrick – MIT CSAIL.

Slides:



Advertisements
Similar presentations
Analysis of Contour Motions Ce Liu William T. Freeman Edward H. Adelson Computer Science and Artificial Intelligence Laboratory Massachusetts Institute.
Advertisements

Visuo-Motor Relationships: Plasticity and Development Read: Rosenbaun Chapters 2, 6 + may.doc.
Learning Techniques for Video Shot Detection Under the guidance of Prof. Sharat Chandran by M. Nithya.
Perception and Perspective in Robotics Paul Fitzpatrick MIT Computer Science and Artificial Intelligence Laboratory Humanoid Robotics Group Goal To build.
The DayOne project: how far can a robot develop in 24 hours? Paul Fitzpatrick MIT CSAIL.
System Integration and Experimental Results Intelligent Robotics Research Centre (IRRC) Department of Electrical and Computer Systems Engineering Monash.
Where has all the data gone? In a complex system such as Metalman, the interaction of various components can generate unwanted dynamics such as dead time.
Yiannis Demiris and Anthony Dearden By James Gilbert.
CS 561, Sessions 27 1 Towards intelligent machines Thanks to CSCI561, we now know how to… - Search (and play games) - Build a knowledge base using FOL.
Highlights Lecture on the image part (10) Automatic Perception 16
A Smart Sensor to Detect the Falls of the Elderly.
Matching Compare region of image to region of image. –We talked about this for stereo. –Important for motion. Epipolar constraint unknown. But motion small.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Cognitive Processes PSY 334 Chapter 2 – Perception.
 For many years human being has been trying to recreate the complex mechanisms that human body forms & to copy or imitate human systems  As a result.
EE392J Final Project, March 20, Multiple Camera Object Tracking Helmy Eltoukhy and Khaled Salama.
Charts and Graphs V
A Review of Children, Humanoid Robots and Caregivers (Arsenio, 2004) COM3240 – Week 3 Presented by Gizdem Akdur.
A Brief Overview of Computer Vision Jinxiang Chai.
The Whole World in Your Hand: Active and Interactive Segmentation The Whole World in Your Hand: Active and Interactive Segmentation – Artur Arsenio, Paul.
Human-Computer Interaction Human-Computer Interaction Tracking Hanyang University Jong-Il Park.
Active Vision Key points: Acting to obtain information Eye movements Depth from motion parallax Extracting motion information from a spatio-temporal pattern.
Page: 1 PHAM VAN Tien Real-Time Approach for Auto-Adjusting Vision System Reading Class International Graduate School of Dynamic Intelligent Systems.
Towards Cognitive Robotics Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Christian.
CSCE 5013 Computer Vision Fall 2011 Prof. John Gauch
Vision & Recognition. From a different direction At different times, particularly if it has been modified in the interval In different light A particular.
University of Amsterdam Search, Navigate, and Actuate - Qualitative Navigation Arnoud Visser 1 Search, Navigate, and Actuate Qualitative Navigation.
Supervised Learning of Edges and Object Boundaries Piotr Dollár Zhuowen Tu Serge Belongie.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
PSEUDO-RELEVANCE FEEDBACK FOR MULTIMEDIA RETRIEVAL Seo Seok Jun.
Natural Tasking of Robots Based on Human Interaction Cues Brian Scassellati, Bryan Adams, Aaron Edsinger, Matthew Marjanovic MIT Artificial Intelligence.
1 Artificial Intelligence: Vision Stages of analysis Low level vision Surfaces and distance Object Matching.
1 Research Question  Can a vision-based mobile robot  with limited computation and memory,  and rapidly varying camera positions,  operate autonomously.
Chapter 5 Multi-Cue 3D Model- Based Object Tracking Geoffrey Taylor Lindsay Kleeman Intelligent Robotics Research Centre (IRRC) Department of Electrical.
Cognitive ability: Challenge: How to recognize objects in a scene; where are the object’s boundaries? This problem is known as ‘image segmentation’ in.
Evaluating Perceptual Cue Reliabilities Robert Jacobs Department of Brain and Cognitive Sciences University of Rochester.
Autonomous Robots Vision © Manfred Huber 2014.
Segmentation of Vehicles in Traffic Video Tun-Yu Chiang Wilson Lau.
Exploiting cross-modal rhythm for robot perception of objects Artur M. Arsenio Paul Fitzpatrick MIT Computer Science and Artificial Intelligence Laboratory.
The Cyber-Physical Bike A Step Toward Safer Green Transportation.
Give examples of the way that virtual reality can be used in Psychology.
DARPA Mobile Autonomous Robot Software BAA99-09 July 1999 Natural Tasking of Robots Based on Human Interaction Cues Cynthia Breazeal Rodney Brooks Brian.
Giorgio Metta · Paul Fitzpatrick Humanoid Robotics Group MIT AI Lab Better Vision through Manipulation Manipulation Manipulation.
Robotics/Machine Vision Robert Love, Venkat Jayaraman July 17, 2008 SSTP Seminar – Lecture 7.
GM-Carnegie Mellon Autonomous Driving CRL 1 TitleRobust Detection of Curbs Thrust AreaPerception Project LeadDavid Wettergreen, CMU Wende Zhang, GM Inna.
Feel the beat: using cross-modal rhythm to integrate perception of objects, others, and self Paul Fitzpatrick and Artur M. Arsenio CSAIL, MIT.
Li Fei-Fei, Stanford Rob Fergus, NYU Antonio Torralba, MIT Recognizing and Learning Object Categories: Year 2009 ICCV 2009 Kyoto, Short Course, September.
AN ACTIVE VISION APPROACH TO OBJECT SEGMENTATION – Paul Fitzpatrick – MIT CSAIL.
Manipulation in Human Environments
Charts & Graphs CTEC V
San Diego May 22, 2013 Giovanni Saponaro Giampiero Salvi
Modeling the development of mirror neurons Problem Solution
Cognitive Processes PSY 334
Manipulation in Human Environments
Properties of human stereo processing
Stereo Vision Applications
New horizons in the artificial vision
Opportunistic Perception
Perception and Perspective in Robotics
Learning about Objects
Paul Fitzpatrick Humanoid Robotics Group MIT AI Lab
ICCV 2009 Kyoto, Short Course, September 24
Visual Vertical Alignments and Skill Integration
Better Vision through Experimental Manipulation
Filtering Things to take away from this lecture An image as a function
Perception.
© 2010 Cengage Learning Engineering. All Rights Reserved.
Filtering An image as a function Digital vs. continuous images
Perception.
Learning complex visual concepts
Presentation transcript:

Object Lesson: Discovering and Learning to Recognize Objects Object Lesson: Discovering and Learning to Recognize Objects – Paul Fitzpatrick – MIT CSAIL USA

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 robots and learning  Robots have access to physics, and physics is a good teacher  Physics won’t let you believe the wrong thing for long  Robot perception should ideally integrate experimentation, or at least learn from (non- fatal) mistakes forward seems safe…

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 a challenge: object perception  Object perception is a key enabling technology  Many components: –Object detection –Object segmentation –Object recognition  Typical systems require human-prepared training data; can we use autonomous experimentation?

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 a challenge: object perception  Object perception is a key enabling technology  Many components: –Object detection –Object segmentation –Object recognition  Typical systems require human-prepared training data; can we use autonomous experimentation? Fruit detection

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 a challenge: object perception  Object perception is a key enabling technology  Many components: –Object detection –Object segmentation –Object recognition  Typical systems require human-prepared training data; can we use autonomous experimentation? Fruit segmentation

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 a challenge: object perception  Object perception is a key enabling technology  Many components: –Object detection –Object segmentation –Object recognition  Typical systems require human-prepared training data – can’t adapt to new situations autonomously Fruit recognition

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 talk overview  Perceiving through experiment –Example: active segmentation  Learning new perceptual abilities opportunistically –Example: detecting edge orientation –Example: object detection, segmentation, recognition  An architecture for opportunistic learning –Example: learning about, and through, search activity

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 talk overview  Perceiving through experiment –Example: active segmentation  Learning new perceptual abilities opportunistically –Example: detecting edge orientation –Example: object detection, segmentation, recognition  An architecture for opportunistic learning –Example: learning about, and through, search activity

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 active segmentation  Object boundaries are not always easy to detect visually  Solution: Cog sweeps arm through ambiguous area  Any resulting object motion helps segmentation  Robot can learn to recognize and segment object without further contact

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 active segmentation  Detect contact between arm and object using fast, coarse processing on optic flow signal  Do detailed comparison of motion immediately before and after collision  Use minimum-cut algorithm to generate best segmentation

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 active segmentation

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 active segmentation  Not always practical!  No good for objects the robot can view but not touch  No good for very big or very small objects  But fine for objects the robot is expected to manipulate Head segmentation the hard way!

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 listening to physics  Active segmentation is useful even if robot normally depends on other segmentation cues (color, stereo)  If passive segmentation is incorrect and robot fails to grasp object, active segmentation can use even clumsy collision to get truth  Seems silly not to use this feedback from physics and keep making the same mistake table car

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 talk overview  Perceiving through experiment –Example: active segmentation  Learning new perceptual abilities opportunistically –Example: detecting edge orientation –Example: object detection, segmentation, recognition  An architecture for opportunistic learning –Example: learning about, and through, search activity

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 talk overview  Perceiving through experiment –Example: active segmentation  Learning new perceptual abilities opportunistically –Example: detecting edge orientation –Example: object detection, segmentation, recognition  An architecture for opportunistic learning –Example: learning about, and through, search activity

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 opportunistic learning  To begin with, Cog has three categories of perceptual abilities: –Judgements it can currently make (e.g. about color, motion, time) –Judgements it can sometimes make (e.g. boundary of object, identity of object) –Judgements it cannot currently make (e.g. counting objects)  With opportunistic learning, the robot takes judgements in the sometimes category and works to promote them to the can category by finding reliable correlated features that are more frequently available –Example: analysis of boundaries detected through motion yields purely visual features that are predictive of edge orientation –Example: assuming a non-hostile environment (some continuity in time and space) segmented views of objects can be grouped and purely visual features inferred that are characteristic of distinct objects

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 training a model of edge appearance  Robot initially only perceives oriented edges through active segmentation procedure  Robot collects samples of edge appearance along boundary, and builds a look- up table from appearance to orientation angle  Now can perceive orientation directly  This is often built in, but it doesn’t have to be

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 most frequent samples 1 st 21 st 41 st 61 st 81 st 101 st 121 st 141 st 161 st 181 st 201 st 221 st

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 some tests Red = horizontal Green = vertical

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 natural images  45 0 

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 talk overview  Perceiving through experiment –Example: active segmentation  Learning new perceptual abilities opportunistically –Example: detecting edge orientation –Example: object detection, segmentation, recognition  An architecture for opportunistic learning –Example: learning about, and through, search activity

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 on to object detection…

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 on to object detection… look for this……in this

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 on to object detection…

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 on to object detection…

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 on to object detection… geometry alone geometry + color

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 other examples

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 other examples

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 other examples

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 just for fun look for this……in this result

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 real object in real images

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 yellow on yellow

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 multiple objects camera image response for each object implicated edges found and grouped

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 attention

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 first time seeing a ball sees ball, “thinks” it is cube correctly differentiates ball and cube robot’s current view recognized object (as seen during poking) pokes, segments ball

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 open object recognition

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 talk overview  Perceiving through experiment –Example: active segmentation  Learning new perceptual abilities opportunistically –Example: detecting edge orientation –Example: object detection, segmentation, recognition  An architecture for opportunistic learning –Example: learning about, and through, search activity

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 talk overview  Perceiving through experiment –Example: active segmentation  Learning new perceptual abilities opportunistically –Example: detecting edge orientation –Example: object detection, segmentation, recognition  An architecture for opportunistic learning –Example: learning about, and through, search activity

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 physically-grounded perception active segmentation object segmentation edge catalog object detection (recognition, localization, contact-free segmentation) affordance exploitation [with Giorgio Metta] (rolling) manipulator detection (robot, human)

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 socially-grounded perception

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 socially-grounded perception

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 opportunistic architecture: a virtuous circle familiar activities familiar entities (objects, actors, properties, …) use constraint of familiar activity to discover unfamiliar entity used within it reveal the structure of unfamiliar activities by tracking familiar entities into and through them

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 a virtuous circle poking, chatting car, ball, cube, and their names discover car, ball, and cube through poking; discover their names through chatting

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 a virtuous circle poking, chatting, search car, ball, cube, and their names follow named objects into search activity, and observe the structure of search

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 learning about search

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 a virtuous circle poking, chatting, search car, ball, cube, and their names follow named objects into search activity, and observe the structure of search

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 a virtuous circle poking, chatting, searching car, ball, cube, toma, and their names discover novel object through poking, learn its name (e.g. ‘toma’) indirectly during search

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 finding the toma

Paul Fitzpatrick, MIT CSAIL, Humanoids 2003 conclusion: why do this?  The quest for truly flexible robots  Humanoid form is general-purpose, mechanically flexible  Robots that really live and work amongst us will need to be as general-purpose and adaptive perceptually as they are mechanically