Probabilistic Robotics

Slides:



Advertisements
Similar presentations
State Estimation and Kalman Filtering CS B659 Spring 2013 Kris Hauser.
Advertisements

SA-1 Probabilistic Robotics Planning and Control: Partially Observable Markov Decision Processes.
Hidden Markov Models Reading: Russell and Norvig, Chapter 15, Sections
Monte Carlo Localization for Mobile Robots Karan M. Gupta 03/10/2004
IR Lab, 16th Oct 2007 Zeyn Saigol
Markov Localization & Bayes Filtering 1 with Kalman Filters Discrete Filters Particle Filters Slides adapted from Thrun et al., Probabilistic Robotics.
Probabilistic Robotics Probabilistic Motion Models.
1 Slides for the book: Probabilistic Robotics Authors: Sebastian Thrun Wolfram Burgard Dieter Fox Publisher: MIT Press, Web site for the book & more.
Bayesian Robot Programming & Probabilistic Robotics Pavel Petrovič Department of Applied Informatics, Faculty of Mathematics, Physics and Informatics
Probabilistic Robotics Probabilistic Motion Models.
Bayes Filters Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics TexPoint fonts used in EMF. Read the.
Recursive Bayes Filtering Advanced AI
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
1.Examples of using probabilistic ideas in robotics 2.Reverend Bayes and review of probabilistic ideas 3.Introduction to Bayesian AI 4.Simple example.
Autonomous Robot Navigation Panos Trahanias ΗΥ475 Fall 2007.
Probabilistic Methods in Mobile Robotics. Stereo cameras Infra-red Sonar Laser range-finder Sonar Tactiles.
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
City College of New York 1 Dr. John (Jizhong) Xiao Department of Electrical Engineering City College of New York A Taste of Localization.
Probability: Review TexPoint fonts used in EMF.
Part 2 of 3: Bayesian Network and Dynamic Bayesian Network.
Monte Carlo Localization
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
City College of New York 1 Dr. John (Jizhong) Xiao Department of Electrical Engineering City College of New York A Taste of Localization.
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotics Research Laboratory University of Southern California
HCI / CprE / ComS 575: Computational Perception
Bayesian Filtering for Robot Localization
Mobile Robot controlled by Kalman Filter
Probability in Robotics
Markov Localization & Bayes Filtering
Localization and Mapping (3)
Lab 4 1.Get an image into a ROS node 2.Find all the orange pixels (suggest HSV) 3.Identify the midpoint of all the orange pixels 4.Explore the findContours.
From Bayesian Filtering to Particle Filters Dieter Fox University of Washington Joint work with W. Burgard, F. Dellaert, C. Kwok, S. Thrun.
SA-1 Mapping with Known Poses Ch 4.2 and Ch 9. 2 Why Mapping? Learning maps is one of the fundamental problems in mobile robotics Maps allow robots to.
1 Robot Motion and Perception (ch. 5, 6) These two chapters discuss how one obtains the motion model and measurement model mentioned before. They are needed.
Visibility Graph. Voronoi Diagram Control is easy: stay equidistant away from closest obstacles.
1 Robot Environment Interaction Environment perception provides information about the environment’s state, and it tends to increase the robot’s knowledge.
TKK | Automation Technology Laboratory Partially Observable Markov Decision Process (Chapter 15 & 16) José Luis Peralta.
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 6.1: Bayes Filter Jürgen Sturm Technische Universität München.
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 5.2: Recap on Probability Theory Jürgen Sturm Technische Universität.
Mobile Robot Localization (ch. 7)
City College of New York 1 Dr. Jizhong Xiao Department of Electrical Engineering City College of New York Advanced Mobile Robotics.
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 5.3: Reasoning with Bayes Law Jürgen Sturm Technische Universität.
4 Proposed Research Projects SmartHome – Encouraging patients with mild cognitive disabilities to use digital memory notebook for activities of daily living.
Robotics Club: 5:30 this evening
Probability in Robotics Trends in Robotics Research Reactive Paradigm (mid-80’s) no models relies heavily on good sensing Probabilistic Robotics (since.
Probabilistic Robotics Introduction.  Robotics is the science of perceiving and manipulating the physical world through computer-controlled devices.
Probability in Robotics Trends in Robotics Research Reactive Paradigm (mid-80’s) no models relies heavily on good sensing Probabilistic Robotics (since.
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
CSE 473 Uncertainty. © UW CSE AI Faculty 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one has stood.
Probabilistic Robotics Introduction. SA-1 2 Introduction  Robotics is the science of perceiving and manipulating the physical world through computer-controlled.
Probabilistic Robotics Probability Theory Basics Error Propagation Slides from Autonomous Robots (Siegwart and Nourbaksh), Chapter 5 Probabilistic Robotics.
Autonomous Mobile Robots Autonomous Systems Lab Zürich Probabilistic Map Based Localization "Position" Global Map PerceptionMotion Control Cognition Real.
General approach: A: action S: pose O: observation Position at time t depends on position previous position and action, and current observation.
Matching ® ® ® Global Map Local Map … … … obstacle Where am I on the global map?                                   
Probabilistic Robotics Bayes Filter Implementations Gaussian filters.
CSE 468/568 Deadlines Lab1 grades out tomorrow (5/1) HW2 grades out by weekend Lab 3 grades out next weekend HW3 – Probability homework out Due 5/7 FINALS:
Review of Probability.
Probabilistic Robotics
Markov ó Kalman Filter Localization
Course: Autonomous Machine Learning
State Estimation Probability, Bayes Filtering
Probability Review 11/22/2018.
Probabilistic Robotics
CSE-490DF Robotics Capstone
A Short Introduction to the Bayes Filter and Related Models
EE-565: Mobile Robotics Non-Parametric Filters Module 2, Lecture 5
Probabilistic Map Based Localization
Probability Review 2/24/2019.
Presentation transcript:

Probabilistic Robotics Advanced Mobile Robotics Probabilistic Robotics Dr. Jizhong Xiao Department of Electrical Engineering City College of New York jxiao@ccny.cuny.edu

Robot Navigation Fundamental problems to provide a mobile robot with autonomous capabilities: Where am I going What’s the best way there? Where have I been?  how to create an environmental map with imperfect sensors? Where am I?  how a robot can tell where it is on a map? What if you’re lost and don’t have a map? Mission Planning Path Planning Mapping Localization Robot SLAM

Representation of the Environment Environment Representation Continuos Metric ® x,y,q Discrete Metric ® metric grid Discrete Topological ® topological grid

Localization, Where am I? Perception Odometry, Dead Reckoning Localization base on external sensors, beacons or landmarks Probabilistic Map Based Localization

Localization Methods Mathematic Background, Bayes Filter Markov Localization: Central idea: represent the robot’s belief by a probability distribution over possible positions, and uses Bayes’ rule and convolution to update the belief whenever the robot senses or moves Markov Assumption: past and future data are independent if one knows the current state Kalman Filtering Central idea: posing localization problem as a sensor fusion problem Assumption: gaussian distribution function Particle Filtering Central idea: Sample-based, nonparametric Filter Monte-Carlo method SLAM (simultaneous localization and mapping) Multi-robot localization

Markov Localization Applying probability theory to robot localization Markov localization uses an explicit, discrete representation for the probability of all position in the state space. This is usually done by representing the environment by a grid or a topological graph with a finite number of possible states (positions). During each update, the probability for each state (element) of the entire space is updated.

Markov Localization Example Assume the robot position is one- dimensional The robot is placed somewhere in the environment but it is not told its location The robot queries its sensors and finds out it is next to a door

Markov Localization Example The robot moves one meter forward. To account for inherent noise in robot motion the new belief is smoother The robot queries its sensors and again it finds itself next to a door

Probabilistic Robotics Falls in between model-based and behavior-based techniques There are models, and sensor measurements, but they are assumed to be incomplete and insufficient for control Statistics provides the mathematical glue to integrate models and sensor measurements Basic Mathematics Probabilities Bayes rule Bayes filters

The next slides are provided by the authors of the book "Probabilistic Robotics“, you can download from the website:  http://www.probabilistic-robotics.org/

Probabilistic Robotics Mathematic Background Probabilities Bayes rule Bayes filters

Probabilistic Robotics Key idea: Explicit representation of uncertainty using the calculus of probability theory Perception = state estimation Action = utility optimization

Axioms of Probability Theory Pr(A) denotes probability that proposition A is true.

A Closer Look at Axiom 3 B

Using the Axioms

Discrete Random Variables X denotes a random variable. X can take on a countable number of values in {x1, x2, …, xn}. P(X=xi), or P(xi), is the probability that the random variable X takes on value xi. P( ) is called probability mass function. E.g. .

Continuous Random Variables X takes on values in the continuum. p(X=x), or p(x), is a probability density function. E.g. p(x) x

Joint and Conditional Probability P(X=x and Y=y) = P(x,y) If X and Y are independent then P(x,y) = P(x) P(y) P(x | y) is the probability of x given y P(x | y) = P(x,y) / P(y) P(x,y) = P(x | y) P(y) If X and Y are independent then P(x | y) = P(x)

Law of Total Probability, Marginals Discrete case Continuous case

Bayes Formula If y is a new sensor reading Prior probability distribution  Posterior probability distribution  Generative model, characteristics of the sensor   Does not depend on x

Normalization Algorithm:

Conditioning Law of total probability:

Bayes Rule with Background Knowledge

Conditioning Total probability:

Conditional Independence equivalent to and

Simple Example of State Estimation Suppose a robot obtains measurement z What is P(open|z)?

Causal vs. Diagnostic Reasoning P(open|z) is diagnostic. P(z|open) is causal. Often causal knowledge is easier to obtain. Bayes rule allows us to use causal knowledge: count frequencies!

Example P(z|open) = 0.6 P(z|open) = 0.3 P(open) = P(open) = 0.5 z raises the probability that the door is open.

Combining Evidence Suppose our robot obtains another observation z2. How can we integrate this new information? More generally, how can we estimate P(x| z1...zn )?

Recursive Bayesian Updating Markov assumption: zn is independent of z1,...,zn-1 if we know x.

Example: Second Measurement P(z2|open) = 0.5 P(z2|open) = 0.6 P(open|z1)=2/3 z2 lowers the probability that the door is open.

Actions Often the world is dynamic since change the world. actions carried out by the robot, actions carried out by other agents, or just the time passing by change the world. How can we incorporate such actions?

Typical Actions The robot turns its wheels to move The robot uses its manipulator to grasp an object Plants grow over time… Actions are never carried out with absolute certainty. In contrast to measurements, actions generally increase the uncertainty.

Modeling Actions To incorporate the outcome of an action u into the current “belief”, we use the conditional pdf P(x|u,x’) This term specifies the pdf that executing u changes the state from x’ to x.

Example: Closing the door

State Transitions P(x|u,x’) for u = “close door”: If the door is open, the action “close door” succeeds in 90% of all cases.

Integrating the Outcome of Actions Continuous case: Discrete case:

Example: The Resulting Belief

Bayes Filters: Framework Given: Stream of observations z and action data u: Sensor model P(z|x). Action model P(x|u,x’). Prior probability of the system state P(x). Wanted: Estimate of the state X of a dynamical system. The posterior of the state is also called Belief:

Markov Assumption Markov Assumption: Measurement probability   State transition probability Markov Assumption: past and future data are independent if one knows the current state Underlying Assumptions Static world, Independent noise Perfect model, no approximation errors

Bayes Filters z = observation u = action x = state Bayes Markov Total prob. Markov Markov

Bayes Filter Algorithm Algorithm Bayes_filter( Bel(x),d ): h=0 If d is a perceptual data item z then For all x do Else if d is an action data item u then Return Bel’(x)

Bayes Filters are Familiar! Kalman filters Particle filters Hidden Markov models Dynamic Bayesian networks Partially Observable Markov Decision Processes (POMDPs)

Summary Bayes rule allows us to compute probabilities that are hard to assess otherwise. Under the Markov assumption, recursive Bayesian updating can be used to efficiently combine evidence. Bayes filters are a probabilistic tool for estimating the state of dynamic systems.

Break Exercises 2 and 3: pp36~37 of textbook