Introduction to Robotics Cognitive Robotics 2007/2008 Period 3 Marco Wiering.

Slides:



Advertisements
Similar presentations
Clustering & image segmentation Goal::Identify groups of pixels that go together Segmentation.
Advertisements

Motion planning, control and obstacle avoidance D. Calisi.
AuRA: Principles and Practice in Review
DESIGN OF A GENERIC PATH PATH PLANNING SYSTEM AILAB Path Planning Workgroup.
Uncertainty Representation. Gaussian Distribution variance Standard deviation.
Artificial Intelligence Intelligent Agents 2 Norvig, Ch. 2 and Nilsson, Ch. 1, 2, 25.
Introduction to Robotics In the name of Allah. Introduction to Robotics o Leila Sharif o o Lecture #2: The Big.
AI Lab Weekly Seminar By: Buluç Çelik.
Autonomous Robot Navigation Panos Trahanias ΗΥ475 Fall 2007.
ECE 7340: Building Intelligent Robots QUALITATIVE NAVIGATION FOR MOBILE ROBOTS Tod S. Levitt Daryl T. Lawton Presented by: Aniket Samant.
Introduction to Robotics
Robotics CSPP Artificial Intelligence March 10, 2004.
ECE 4340/7340 Exam #2 Review Winter Sensing and Perception CMUcam and image representation (RGB, YUV) Percept; logical sensors Logical redundancy.
Brent Dingle Marco A. Morales Texas A&M University, Spring 2002
Chapter 25: Robotics April 27, The Week Ahead … Wednesday: Dmitrii Zagorodnov Thursday: Jeff Elser’s presentation, general discussion Friday: Rafal.
Vision January 10, Today's Agenda ● Some general notes on vision ● Colorspaces ● Numbers and Java ● Feature detection ● Rigid body motion.
Introduction to mobile robots Slides modified from Maja Mataric’s CSCI445, USC.
Chapter 5: Path Planning Hadi Moradi. Motivation Need to choose a path for the end effector that avoids collisions and singularities Collisions are easy.
Planetary Surface Robotics ENAE 788U, Spring 2005 U N I V E R S I T Y O F MARYLAND Lecture 8 Mapping 5 April, 2005.
Simultaneous Localization and Map Building System for Prototype Mars Rover CECS 398 Capstone Design I October 24, 2001.
Behavior- Based Approaches Behavior- Based Approaches.
Robotics In which agents are endowed with physical effectors with which to do mischief.
N.B. Please register for the course with the ITO Please attend a practical in the coming week: –Either 10:00, 13:05 Tuesday –Or 10:00, 13:05 Friday If.
Quick Overview of Robotics and Computer Vision. Computer Vision Agent Environment camera Light ?
June 12, 2001 Jeong-Su Han An Autonomous Vehicle for People with Motor Disabilities by G. Bourhis, O.Horn, O.Habert and A. Pruski Paper Review.
Mohammed Rizwan Adil, Chidambaram Alagappan., and Swathi Dumpala Basaveswara.
Fuzzy control of a mobile robot Implementation using a MATLAB-based rapid prototyping system.
Robotica Lezione 1. Robotica - Lecture 12 Objectives - I General aspects of robotics –Situated Agents –Autonomous Vehicles –Dynamical Agents Implementing.
Lab 3 How’d it go?.
9/14/2015CS225B Kurt Konolige Locomotion of Wheeled Robots 3 wheels are sufficient and guarantee stability Differential drive (TurtleBot) Car drive (Ackerman.
DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; March Adaptive Intelligent Mobile Robotics Leslie Pack Kaelbling Artificial Intelligence.
Robotica Lecture 3. 2 Robot Control Robot control is the mean by which the sensing and action of a robot are coordinated The infinitely many possible.
AIBO Programming AIBO Programming Introduction to Mobile Robotics AIBO Programming course 2005/2006 Period 1 Frank Dignum & Marco Wiering.
SPIE'01CIRL-JHU1 Dynamic Composition of Tracking Primitives for Interactive Vision-Guided Navigation D. Burschka and G. Hager Computational Interaction.
Lecture 2: Introduction to Concepts in Robotics
CS654: Digital Image Analysis Lecture 3: Data Structure for Image Analysis.
Towards Cognitive Robotics Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Christian.
Robotica Lecture 3. 2 Robot Control Robot control is the mean by which the sensing and action of a robot are coordinated The infinitely many possible.
University of Amsterdam Search, Navigate, and Actuate - Qualitative Navigation Arnoud Visser 1 Search, Navigate, and Actuate Qualitative Navigation.
Computer Vision 776 Jan-Michael Frahm 12/05/2011 Many slides from Derek Hoiem, James Hays.
The Hardware Design of the Humanoid Robot RO-PE and the Self-localization Algorithm in RoboCup Tian Bo Control and Mechatronics Lab Mechanical Engineering.
1 Research Question  Can a vision-based mobile robot  with limited computation and memory,  and rapidly varying camera positions,  operate autonomously.
Introduction to Artificial Intelligence CS 438 Spring 2008 Today –AIMA, Ch. 25 –Robotics Thursday –Robotics continued Home Work due next Tuesday –Ch. 13:
Chapter 7. Learning through Imitation and Exploration: Towards Humanoid Robots that Learn from Humans in Creating Brain-like Intelligence. Course: Robots.
Instructional Objective  Define an agent  Define an Intelligent agent  Define a Rational agent  Discuss different types of environment  Explain classes.
Robotics Club: 5:30 this evening
Range-Only SLAM for Robots Operating Cooperatively with Sensor Networks Authors: Joseph Djugash, Sanjiv Singh, George Kantor and Wei Zhang Reading Assignment.
Robotics/Machine Vision Robert Love, Venkat Jayaraman July 17, 2008 SSTP Seminar – Lecture 7.
Path Planning Based on Ant Colony Algorithm and Distributed Local Navigation for Multi-Robot Systems International Conference on Mechatronics and Automation.
Robotics Chapter 6 – Machine Vision Dr. Amit Goradia.
Vision-Guided Humanoid Footstep Planning for Dynamic Environments P. Michel, J. Chestnutt, J. Kuffner, T. Kanade Carnegie Mellon University – Robotics.
Basilio Bona DAUIN – Politecnico di Torino
Autonomous Mobile Robots Autonomous Systems Lab Zürich Probabilistic Map Based Localization "Position" Global Map PerceptionMotion Control Cognition Real.
3D Perception and Environment Map Generation for Humanoid Robot Navigation A DISCUSSION OF: -BY ANGELA FILLEY.
CSPP Artificial Intelligence March 10, 2004
Vision-Guided Humanoid Footstep Planning for Dynamic Environments
CMSC Artificial Intelligence March 11, 2008
Paper – Stephen Se, David Lowe, Jim Little
CS b659: Intelligent Robotics
Introduction to Robots
Schedule for next 2 weeks
Cover Option2.
Vehicle Segmentation and Tracking in the Presence of Occlusions
Locomotion of Wheeled Robots
Multiple View Geometry for Robotics
Probabilistic Map Based Localization
Robot Intelligence Kevin Warwick.
Chapter 4 . Trajectory planning and Inverse kinematics
Presentation transcript:

Introduction to Robotics Cognitive Robotics 2007/2008 Period 3 Marco Wiering

Cognitive Robotics Introduction to Robotics terminology

Cognitive Robotics Robots and their environments A robot is an active, mobile, physical, artificial agent whose environment is the real physical world We will concentrate on Autonomous robots that choose their own actions based on their perception of the world The problem of making robot controllers is that the real world is very demanding: The world cannot be completely perceived, it is non-deterministic, it is dynamical, and continuous

Cognitive Robotics Applications of robots Manufacturing (repeatable tasks on production belt; car and micro-electronics industry) Construction industry (e.g., for shaving sheep, picking tomatoes) Delivery and security robots Unmanned vehicles (cars, airplanes, underwater) Tasks in dangerous environments (earthquakes, chemical substances, radio-active environments) Space exploration missions Household tasks (vacuum cleaner, lawnmower) Entertainment industry

Cognitive Robotics The Syllabus Sensors Vision Actuators Motion (Forward/Inverse) Kinematics Drift Localization Navigation Basic behaviors Complex behaviors Multi-robot behaviors Inertia Torque Compass Joint Bumpers Landmark Geometric Map.....

Cognitive Robotics Types of robots (I) Static Robots vs Mobile Robots Lunokhod

Cognitive Robotics Types of robots (II) Wheeled Robots VS Legged Robots Sojourner (NASA)

Cognitive Robotics Types of robots (III) Robots, – Small, cheap robots, – cheap sensors (no sonar or laser) Microbots, Nanobots

Cognitive Robotics Levels of abstraction Action Actuato rs Percepti on External World Sensor s Cognitio n PHYSICAL/HARDWARE LEVEL DEVICE LEVEL COMPUTATIONAL LEVEL Sensor Drivers Or Sensing Libraries Comm HW Actuator Drivers Or Motion Libraries Communication Interface

Cognitive Robotics Intelligent Robot Tasks Perception –sensing, modelling of the world –Communication (listening) Cognition –behaviours, action selection, planning, learning –multi-robot coordination, teamwork –response to opponent, multi-agent learning Action –motion, navigation, obstacle avoidance –Communication (telling)

Cognitive Robotics Perception Non-visual sensors Vision (segmentation, colour) Localization

Cognitive Robotics Perception: Non-visual sensors Laser range finder (scanner) Sonar (SOund NAvigation and Ranging) Proprioception (what are the joint-positions?) Odometry (measures wheel rotations) Force Sensors Touch Sensors (exist also in an AIBO) Infrared sensors

Cognitive Robotics Perception: Vision Vision is a way to relate measurement to scene structure –Our human environments are shaped to be navigated by vision E.g., road lines –Problem: vision technology is not very well-developed Recognition of shapes, forms, …. –Solution: in most of cases, we don’t need full recognition Use our knowledge of the domain to ease vision –E.g.: Green space in a soccer field means free (void) space. Two kinds of vision: –Passive vision: static cameras Processing of snapshots –Active vision: the camera moves. Intricate relation between camera and the environment

Cognitive Robotics Perception: Vision Active Vision Important geometric relation between the camera and the environment –Movement of the camera should produce an (expected) change in the image Useful to increase the visual information of an item –Move to avoid another object that blocks the vision –Move to have another viewpoint of the object, and ease recognition –Move to measure distances by comparison of images An improvement: Stereo Vision Active Vision is highly sensible to calibration –Geometric calibration –Color calibration

Cognitive Robotics Perception: Vision Active Vision Color calibration –Identify the colors of landmarks and important objects –Adaptation to local light condition –Saturation of color Colored blobs identified as objects Problem: threshold selection Geometric calibration –Position of the camera related to the floor –At least 3 coordinate systems Egocentric coordinates Camera coordinates –Translation matrix counting intermediate joints

Cognitive Robotics Perception: Vision Image Segmentation Sort pixels into classes Obstacle: –Red robot –Blue robot –White wall –Yellow goal –Cyan goal –Unknown color Free space : –Green field Undefined occupancy: –Orange ball –White line

Cognitive Robotics Start with a single pixel p and wish to expand from that seed pixel to fill a coherent region. Define a similarity measure S(i, j) such that it produces a high result if pixels i and j are similar Add pixel q to neighbouring pixel p ’s region iff S(p, q) > T for some threshold T. We can then proceed to the other neighbors of p and do likewise, and then those of q. Problems: –highly sensible to the selection of the seed and the Threshold. –computationally expensive because the merging process starts from small initial regions (individual points). Perception: Vision Image Segmentation by region growing (Example by L. Saad and C. Bordenade)

Cognitive Robotics SplitSplit the image. Start by considering the entire image as one region. –If the entire region is coherent (i.e., if all pixels in the region have sufficient similarity), leave it unmodified. –If the region is not sufficiently coherent, split it into four quadrants and recursively apply these steps to each new region. The “splitting” phase builds a quadtree –several adjacent squares of varying sizes might have similar characteristics. MergeMerge these squares into larger coherent regions from the bottom up. –Since it starts with regions (hopefully) larger than single pixels, this method is more efficient. Perception: Vision Image Segmentation by Split and Merge (Example by C. Urdiales)

Cognitive Robotics Perception: Localization Where am I? Given a map, determine the robot’s location –Landmark locations are known, but the robot’s position is not –From sensor readings, the robot must be able to infer its most likely position on the field –Example : where are the AIBOs on the soccer field?

Cognitive Robotics Scanning Image for Objects Scanlines projected from origin for egocentric coordinates in 5 degree increments Top view of robot Scanlines projected onto RLE image

Cognitive Robotics Measuring Distances with the AIBO’s Camera Assume a common ground plane Assume objects are on the ground plane –Elevated objects will appear further away –Increased distance causes loss of resolution

Cognitive Robotics Identifying Objects in Image Along each scanline: –Identify continuous line of object colors –Filter out noise pixels –Identify colors to form pixel group

Cognitive Robotics Bayesian Filter Why should you care? –Robot and environmental state estimation is a fundamental problem! Nearly all algorithms that exist for spatial reasoning make use of this approach –If you’re working in mobile robotics, you’ll see it over and over! –Very important to understand and appreciate Efficient state estimator –Recursively compute the robot’s current state based on the previous state of the robot What is the robot’s state?

Cognitive Robotics Perception: Localization with Uncertainty Initial state detects nothing: Moves and detects landmark: Moves and detects nothing: Moves and detects landmark:

Cognitive Robotics Planning and Motion Task Planning Motion Planning and Navigation Mapping Motion Planning with Uncertainty (Probabilistic Robotics)

Cognitive Robotics Intelligent Complete Robot Action Actuato rs Percepti on External World Sensor s Cognitio n

Cognitive Robotics Task Planning: Behavior selection ScoreSearchApproachRecover not see ball next to ball see ballnot see ball timeout see ballnot next to ball

Cognitive Robotics Action: Motion Four-legged walking (several joints with degrees of liberty) Head motion (2 joints, 3 degrees of liberty) How to generate complex behaviors (turning, kicking?) Kinematics: relation between the control inputs and the robot motion –Forward kinematics problem Given the control inputs, how does the robot move –Inverse kinematics problem Given a desired motion, which control inputs to choose

Cognitive Robotics Robot Motion A 51-parameter structure is used to specify the gait of the robot. Global Parameters: Height of Body (1) Angle of Body (1) Hop Amplitude (1) Sway Amplitude (1) Walk Period (1) Height of Legs (2) Leg Parameters: Neutral Kinematic Position (3x4) Lifting Velocity (3x4) Lift Time (1x4) Set Down Velocity (3x4) Set Down Time (1x4)

Cognitive Robotics

Cognitive Robotics Approaches for Parameter Setting Trial and error –Tedious, but controlled, and provides knowledge of parameters Search –Large parameter space, local vs. global optima Adaptation –Controlled change by feedback

Cognitive Robotics Forward Kinematics Determines position in space based on joint configuration

Cognitive Robotics Inverse Kinematics Compute joint configuration to attain a desired position and orientation of end effector (tool) of robot More complex than forward kinematics Often also involves path-planning to attain joint configuration from the current one Usually solved algebraically or geometrically, but needs model of the robot! Possibly no solution, one solution, or multiple solutions

Cognitive Robotics An example Let's assume l1 = l2 What is the configuration of the joint-angles if the end-effector is located at (l1, l2)?

Cognitive Robotics World Models (I) Representations of the environment are usually built by means of: In our architecture we use both representations Metric maps Metric maps: explicitly reproduce the metrical structure of the domain good for location, hard for planning Evidence grids e.g., Evidence grids Topological maps Topological maps: represent the environment as a set of meaningful regions. good for planning, hard for location

Cognitive Robotics World Models (II) Topological Map Extraction Metric map thresholding(a) Metric map thresholding –cell occupancy values Hierarchical split(b) Hierarchical split –piramidal cell structure Interlevel merging(c) Interlevel merging –homogeneous cells fusion Intralevel merging(d) Intralevel merging –homogeneous cell classification

Cognitive Robotics Navigation (I) NavigationNavigation consists of finding and tracking a safe path from a departure point to a goal. architecturesNavigation architectures belong to three broad categories: deliberative, reactive and hybrid.

Cognitive Robotics Navigation (II) Deliberative schemesDeliberative schemes require extensive world knowledge to build high-level plans sense-model-plan-act –Usually they use the sense-model-plan-act cycle –problem 1: inability to react rapidly –problem 2: not suitable for (partially) unknown environments. Reactive schemesReactive schemes try to couple sensors and actuators to achieve a fast response. –Easily combine several sensors and goals, –problem 1: the emergent behaviour may be unpredictable –problem 2: the emergent behaviour may be inefficient (prone to fall in local traps). Hybrid schemasHybrid schemas get the best of both approaches.

Cognitive Robotics Reactive system using Potential Fields Create repulsion force around obstacles plus attraction force to the goal

Cognitive Robotics The hybrid architecture

Cognitive Robotics Discussion Robot Controllers need: 1) Perception of the environment; classifying and measuring distances to objects 2) Self localization (dealing with uncertainty) 3) Obstacle-free navigation; path-planning, reactive behaviour 4) Task planning; what is the current/next goal? For multiple-robots the robots also need to: 1) Communicate (e.g. shared world models) 2) Negotiate (goal division among team members)