Download presentation
1
Simultaneous Localization and Mapping
SLAM Simultaneous Localization and Mapping
2
Reinforcement learning to combine different map representations
Occupancy Grid Feature Map Localization Particle filters FastSLAM Reinforcement learning to combine different map representations
3
Occupancy grid / grid map
Simple black-white picture Good for dense places
4
Feature map Good for sparse places
5
Localization Map is known sensors data and robots kinematics is known
Determine the position
6
Localization Discrete time ๐ โ landmarks position
๐ ๐ก - robots position ๐ข ๐ก - control ๐ง ๐ก - sensor information ๐ 1 ๐ข 1 ๐ง 1 โ ๐ 2 ๐ 2 ๐ข 2 ๐ง 2 โ ๐ 3
7
Particle filter requirements
Motion model ๐ ๐ ๐ก ๐ข ๐ก , ๐ ๐กโ1 ) If current position is (๐ฅ,๐ฆ) and the robot movement is ๐๐ฅ, ๐๐ฆ new coordinates are (๐ฅ+๐๐ฅ, ๐ฆ+๐๐ฆ) + noice Usually the noise is Gaussian
8
Particle filter requirements
Measurement model ๐ ๐ง ๐ก ๐ ๐ก , ๐, ๐ ๐ก ) ๐ โ collection of landmark position ๐ 1 , ๐ 2 , โฆ ๐ ๐พ ๐ ๐ก - landmark observed at time ๐ก In simple case each landmark is uniquely identifiable
9
Particle filter We have N particles
Each particle is simply current position For each particle: Update its position using motion model Assign a weight using measurement model Normalize importance weights such that their sum is 1 Resample N particles with probabilities proportional to the weight
10
Particle filter code
11
SLAM In SLAM problem we try to build a map. Most common methods:
Kalman filters (Normal distribution in high-dimensional space) Particle filter (what a particle represents here?)
12
FastSLAM We try to determine robot and landmarks locations based on control and sensor data N particles Robot position Gaussian distribution for each of K landmarks Time complexity O ๐ log ๐พ Space complexity - ?
13
FastSLAM If we know the path ( ๐ 1 โฆ ๐ ๐ก ) ๐ 1 and ๐ 2 are independent
14
FastSLAM ๐ ๐ ๐ก , ๐ ๐ง ๐ก , ๐ข ๐ก , ๐ ๐ก )=๐ ๐ ๐ก ๐ง ๐ก , ๐ข ๐ก , ๐ ๐ก )โ ๐ ๐ ๐ ๐ ๐ก , ๐ง ๐ก , ๐ข ๐ก , ๐ ๐ก ) We have K+1 problems: Estimation of the path Estimation of landmarks location made using Kalman filter.
15
FastSLAM Weights calculation:
Position of a landmark is modeled by Gaussian ๐ค ๐ก ~ ๐ ๐ ๐ก ๐ง ๐ก , ๐ข ๐ก , ๐ ๐ก ) ๐ ๐ ๐ก ๐ง ๐กโ1 , ๐ข ๐ก , ๐ ๐กโ1 ) ~ ๐ ๐ง ๐ก ๐ ๐ ๐ก , ๐ ๐ก , ๐ ๐ก )๐( ๐ ๐ ๐ก ) ๐ ๐ ๐ ๐ก
16
FastSLAM FastSLAM saves landmark positions in a balanced binary tree.
Size of the tree is ๐ ๐พ Sampled particle differs from the previous one in only one leaf.
17
FastSLAM We just create new tree on top of the previous one.
Complexity ๐ log ๐พ Video 2
18
Combining different map representation
There are many ways how we represent a map How we can combine them? Grid map Feature map
19
Model selection Map parameters: Observation likelihood
For given particle we get likelihood of laser observation Average for all particles ๐ผ = 1 ๐ โ๐( ๐ง ๐ก | ๐ ๐ก , ๐ ๐ก , ๐ ๐ก ) Between 0 and 1, large values mean good map ๐ ๐๐๐ - effective sample size ๐ ๐๐๐ = 1 โ ๐ค 2 , here we assume that โ๐ค=1 It is a measure of variance in weight. Suppose all weights are the same, what is ๐ ๐๐๐ ?
20
Reinforcement learning for model selection
SARSA (State-Action-Reward-State-Action) Actions: { ๐ ๐ , ๐ ๐ } โ use grid map of feature map States S = ๐ผ ร ๐ ๐๐๐ ๐ < ๐ ๐๐๐ ๐ ร{๐๐๐๐ก๐ข๐๐ ๐๐๐ก๐๐๐ก๐๐} ๐ผ is divided into 7 intervals ( ) Feature detected โ determines weather a feature was detected on current step. 7ร2ร2=28 states
21
Reinforcement learning for model selection
Reward: For simulations correct robot position is known. Deviation from the correct position gives negative reward. ๐-Greedy, ๐=0.6 Learning rate ๐ผ=0.001 Discounting factor ๐พ=0.9
22
The algorithm
23
The algorithm
24
Results
25
Multi-robot SLAM If the environment is large using only one robot is not enough Centralized approach โ the map is merged than the entire environment is explored Decentralized approach โ robots merge their maps than they meet each other
26
Multi-robot SLAM We need to transform frame of references.
27
Reinforcement learning for model selection
Two robots meat each other and decide how they share their information Actions ๐ ๐๐ - donโt merge maps ๐ ๐๐ - merge with simple transformation matrix ๐ ๐๐ โ use grid-based heuristic to improve transformation matrix ๐ ๐๐ - use feature-based heuristic
28
Reinforcement learning for model selection
States ๐ผ ๐ , ๐ผ ๐ , ๐ ๐๐๐ ๐ < ๐ 2 , ๐ ๐๐๐ ๐ < ๐ 2 3ร3ร2ร2=36 states ๐ผ ๐ - confidence for the transformation matrix for grid- bases method, 3 intervals ( )
29
Reinforcement learning for model selection
Reward For simulations correct robot position is known โ we can get cumulative error for robot position ๐= ๐ธ ๐๐ข๐ ๐ โ ๐ธ ๐๐ข๐ ๐ธ ๐๐ข๐ ๐ - average cumulative error achieved by several runs where the robots immediately merge. ๐- Greedy policy ๐=0.1 ๐ผ=0.001 ๐พ=0.9
30
Results
32
References nar%20Day/Autonomous%20vehicles/theses/dinnissen -thesis.pdf ?via=ihub personal.acfr.usyd.edu.au/nebot/publications/slam/IJRR _slam.htm
33
Questions
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.