Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006.

Slides:



Advertisements
Similar presentations
Lecture 7: Potential Fields and Model Predictive Control
Advertisements

AI Pathfinding Representing the Search Space
P3 / 2004 Register Allocation. Kostis Sagonas 2 Spring 2004 Outline What is register allocation Webs Interference Graphs Graph coloring Spilling Live-Range.
A Hierarchical Multiple Target Tracking Algorithm for Sensor Networks Songhwai Oh and Shankar Sastry EECS, Berkeley Nest Retreat, Jan
Handling Deadlocks n definition, wait-for graphs n fundamental causes of deadlocks n resource allocation graphs and conditions for deadlock existence n.
Resource Management §A resource can be a logical, such as a shared file, or physical, such as a CPU (a node of the distributed system). One of the functions.
Problem Solving by Searching Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 3 Spring 2007.
1 Greedy Forwarding in Dynamic Scale-Free Networks Embedded in Hyperbolic Metric Spaces Dmitri Krioukov CAIDA/UCSD Joint work with F. Papadopoulos, M.
Mobile Robotics Laboratory Institute of Systems and Robotics ISR – Coimbra Aim –Take advantage from intelligent cooperation between mobile robots, so as.
Yu Stephanie Sun 1, Lei Xie 1, Qi Alfred Chen 2, Sanglu Lu 1, Daoxu Chen 1 1 State Key Laboratory for Novel Software Technology, Nanjing University, China.
Generated Waypoint Efficiency: The efficiency considered here is defined as follows: As can be seen from the graph, for the obstruction radius values (200,
DESIGN OF A GENERIC PATH PATH PLANNING SYSTEM AILAB Path Planning Workgroup.
1 E190Q – Project Introduction Autonomous Robot Navigation Team Member 1 Name Team Member 2 Name.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
Game-Theoretic Approaches to Multi-Agent Systems Bernhard Nebel.
Cognitive Colonization The Robotics Institute Carnegie Mellon University Bernardine Dias, Bruce Digney, Martial Hebert, Bart Nabbe, Tony Stentz, Scott.
Decentralized prioritized planning in large multirobot teams Prasanna Velagapudi Paul Scerri Katia Sycara Carnegie Mellon University, Robotics Institute.
Localized Techniques for Power Minimization and Information Gathering in Sensor Networks EE249 Final Presentation David Tong Nguyen Abhijit Davare Mentor:
A Heuristic Bidding Strategy for Multiple Heterogeneous Auctions Patricia Anthony & Nicholas R. Jennings Dept. of Electronics and Computer Science University.
Brent Dingle Marco A. Morales Texas A&M University, Spring 2002
A Free Market Architecture for Distributed Control of a Multirobot System The Robotics Institute Carnegie Mellon University M. Bernardine Dias Tony Stentz.
Optimizing Schedules for Prioritized Path Planning of Multi-Robot Systems Maren Bennewitz Wolfram Burgard Sebastian Thrun.
Randomized Planning for Short Inspection Paths Tim Danner Lydia E. Kavraki Department of Computer Science Rice University.
Robot Motion Planning Bug 2 Probabilistic Roadmaps Bug 2 Probabilistic Roadmaps.
Online Data Gathering for Maximizing Network Lifetime in Sensor Networks IEEE transactions on Mobile Computing Weifa Liang, YuZhen Liu.
Opportunistic Optimization for Market-Based Multirobot Control M. Bernardine Dias and Anthony Stentz Presented by: Wenjin Zhou.
DAMN : A Distributed Architecture for Mobile Navigation Julio K. Rosenblatt Presented By: Chris Miles.
Chapter 5.4 Artificial Intelligence: Pathfinding.
1 Algorithms for Bandwidth Efficient Multicast Routing in Multi-channel Multi-radio Wireless Mesh Networks Hoang Lan Nguyen and Uyen Trang Nguyen Presenter:
June 12, 2001 Jeong-Su Han An Autonomous Vehicle for People with Motor Disabilities by G. Bourhis, O.Horn, O.Habert and A. Pruski Paper Review.
Particle Swarm Optimization Algorithms
Chapter 5.4 Artificial Intelligence: Pathfinding.
Lab 3 How’d it go?.
IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS 2007 (TPDS 2007)
1 Constant Following Distance Simulations CS547 Final Project December 6, 1999 Jeremy Elson.
Multi-Robot Systems. Why Multiple Robots? Some tasks require a team Robotic soccer Some tasks can be decomposed and divided for efficiency Mapping a large.
Mutual Exclusion in Wireless Sensor and Actor Networks IEEE SECON 2006 Ramanuja Vedantham, Zhenyun Zhuang and Raghupathy Sivakumar Presented.
Probabilistic Roadmaps for Path Planning in High-Dimensional Configuration Spaces (1996) L. Kavraki, P. Švestka, J.-C. Latombe, M. Overmars.
Multi-Robot Systems.
Load-Balancing Routing in Multichannel Hybrid Wireless Networks With Single Network Interface So, J.; Vaidya, N. H.; Vehicular Technology, IEEE Transactions.
Optimization of Wavelength Assignment for QoS Multicast in WDM Networks Xiao-Hua Jia, Ding-Zhu Du, Xiao-Dong Hu, Man-Kei Lee, and Jun Gu, IEEE TRANSACTIONS.
Wandering Standpoint Algorithm. Wandering Standpoint Algorithm for local path planning Description: –Local path planning algorithm. Required: –Local distance.
1 Distributed and Optimal Motion Planning for Multiple Mobile Robots Yi Guo and Lynne Parker Center for Engineering Science Advanced Research Computer.
2007/03/26OPLAB, NTUIM1 A Proactive Tree Recovery Mechanism for Resilient Overlay Network Networking, IEEE/ACM Transactions on Volume 15, Issue 1, Feb.
Mobile Agent Migration Problem Yingyue Xu. Energy efficiency requirement of sensor networks Mobile agent computing paradigm Data fusion, distributed processing.
A Passive Approach to Sensor Network Localization Rahul Biswas and Sebastian Thrun International Conference on Intelligent Robots and Systems 2004 Presented.
Algorithmic, Game-theoretic and Logical Foundations
Robotics Club: 5:30 this evening
Behavior-based Multirobot Architectures. Why Behavior Based Control for Multi-Robot Teams? Multi-Robot control naturally grew out of single robot control.
Optimization Problems
Overivew Occupancy Grids -Sonar Models -Bayesian Updating
1 Energy-Efficient Mobile Robot Exploration Yongguo Mei, Yung-Hsiang Lu Purdue University ICRA06.
Artificial Intelligence in Game Design Influence Maps and Decision Making.
Fast SLAM Simultaneous Localization And Mapping using Particle Filter A geometric approach (as opposed to discretization approach)‏ Subhrajit Bhattacharya.
Path Planning Based on Ant Colony Algorithm and Distributed Local Navigation for Multi-Robot Systems International Conference on Mechatronics and Automation.
Learning for Physically Diverse Robot Teams Robot Teams - Chapter 7 CS8803 Autonomous Multi-Robot Systems 10/3/02.
Efficient Placement and Dispatch of Sensors in a Wireless Sensor Network You-Chiun Wang, Chun-Chi Hu, and Yu-Chee Tseng IEEE Transactions on Mobile Computing.
R. Brafman and M. Tennenholtz Presented by Daniel Rasmussen.
Heterogeneous Teams of Modular Robots for Mapping and Exploration by Grabowski et. al.
Chapter 5.4 Artificial Intelligence: Pathfinding
COGNITIVE APPROACH TO ROBOT SPATIAL MAPPING
CS b659: Intelligent Robotics
Schedule for next 2 weeks
Planar Graphs & Euler’s Formula
E190Q – Project Introduction Autonomous Robot Navigation
Finding Heuristics Using Abstraction
Navigation In Dynamic Environment
Market-based Dynamic Task Allocation in Mobile Surveillance Systems
Stefan Oßwald, Philipp Karkowski, Maren Bennewitz
Presentation transcript:

Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Outline Why multiple robots Design issues Basic approaches –Distributed –Centralized –Market-based

Why Multiple Robots Some tasks require a robot team Have potential to finish tasks faster Increase robustness w/ redundancy Compensate sensor uncertainty by merging overlapping information Multiple robots allow for more varied and creative solutions

A Good Multi-Robot System Is: Robust: no single point of failure Optimized, even under dynamic conditions Quick to respond to changes Able to deal with imperfect communication Able to avoid robot interference Able to allocate limited resources Heterogeneous and able to make use of different robot skills

Basic Approaches Distributed –Every robot goes for itself Centralized –Globally coordinate all robots Market-based –Analogy To Real Economy

Distributed Methods Planning responsibility spread over team Each robot basically act independently Robots use locally observable information to coordinate and make their plans

Example: Frontier-Based Exploration Using Multiple Robots (Yamauchi 1998) A highly distributed approach Simple idea: To gain the most new information about the world, move to the boundary between open space and uncertainty territory Frontiers are the boundaries between open space and unexplored space

Occupancy Grid World is represented as grid Each cell in the grid is assigned with a probability of being already occupied/observed The initial probability is all set to.5 Cell status can be Open ( 0.5) Bayesian rule is used to update cells by merging information from each sensor reading (sonar)

Frontier Detection Frontier = Boundary between open and unexplored space. Any open cell adjacent to unknown cell is frontier edge cell. Frontier cells grouped into frontier regions based on adjacency. Accessible frontier = Robot can pass through opening. Inaccessible frontier = Robot cannot pass through opening.

Multi-Robot Navigation Simple algorithm: Each robot goes along the shortest obstacle free path to a frontier region Robots share a common map: All information obtained by any robot is available to all robots Robots are planning path independently Use reactive strategy to avoid collisions Robots may waste time for the same frontiers

An Exploration Sequence

Distributed Methods: Pros & Cons Pros –Very robust. No single point failure –Fast response to dynamic conditions –Little or no communication is required –Easy ….Little computation required Cons –Plans only based on local information –Solutions are often sub-optimal

Centralized Methods Robot team treated as a single “system” with many degrees of freedom A single robot or computer is the “leader” Leader plans optimal tasks for groups Group members send information to leader and carry out actions

Example: Arena (Jia 2004) Robots share a common map and only communicate with a leader Robots compete for resources by their efficiency leader greedily assigns the most efficient tasks Leader coordinate robots to handle interference

Background World representation –Occupancy grid Cost unit –Moving forward one step = Turning 45 degrees Cost overflow –Similar to minimum cost spanning tree –Easy to compute the shortest path –Easy to handle obstacle

Cost Overflow Cost of 45° turning = Cost of one cell’s step Direction priority

Goal Candidates Detection A goal point P should satisfy i.P is passable (M ark the cells in warning range or obstacles/Wall/Unknown cells as impassable ) ii.Some unexplored cells lie in the circle with P as the center and (R + K) as the radium, where R is the warning radius and K is usually 1 Robot cell paths cell observation cell candidate goal

Goal Resource Reserved goal candidates –Robots obtained by competition Recessive goal candidates –The goal points in a given range to a reserved goal point –This distance can be adjusted Goal candidates Recessive goals candidates

Path Resource Path resource is a time-space term For a given time, the cells close to any robot are marked off for safety Looks just like a widened path Basically a reactive strategy goal path resource

Revenue and Utility Revenue –The expected gain of information that robots observe at a goal point Utility used by many other approaches –Utility = revenue – cost Utility in this paper –Utility = Revenue / Cost –Better connected to purpose of smallest cost –No need to care about unit conversion

Greedy Goal Selection Try to maximize the global utility Coordination: robots obtain goal and path resources exclusively Competition: repetitively select the pair of free agent and goal with highest utility Sub-optimal

Simple Algorithm Repeat until map is complete –Repeat #free robots times 1.Cost computation (Also make sure no interference with the busy robots) 2.Select the highest utility task (Compete) 3.Mark off the associated robot and goal points, and nearby goal points

1 st Competition: Interval = 3Competitor:

1 st Competition Result: Interval = 3Competitor:

2 nd Competition Interval = 3Competitor:Satisfied:

2 nd Competition Result Interval = 3Competitor:Satisfied:

3 rd Competition Competitor:Satisfied:Interval =

3 rd Competition Result Competitor:Satisfied:Interval =

Planning Issues Do not transfer a reserved goal point to another free agent (unless necessary). Frequent change of tasks can cause localization error. Quit an assigned task when the goal point is unexpectedly observed by other robots Schedule at most one task for each agent

Possible Variations Still keep busy agents in competition. Remove the goal resources they win from competition. –This prevents those goal resources being assigned to other agents –It is too early to burden a new task on a robot who has not achieved it current task No need to schedule them. –New resources probably will be found when they reach the goals

Handling Failure of Planning It may fail to plan safe paths –When some robot get to a place where it is almost too close to other robot it has no good space to detour –And it choose to just wait there for other robots to move away, which is not known by other robots Avoidance of unexpected obstacle –Robots have simple reactive mechanism –Release resources and try to gain new task

Fail to plan safe paths Competitor:Satisfied:Interval = collision

Reactive Mechanism Competitor:Satisfied:Interval = 3

Exchange Tasks Competitor:Satisfied:Interval =

Some Statistics

Demo

Centralized Methods : Pros Leader can take all relevant information into account for planning Optimal s islution possible! One can try different approximate solutions to this problem

Centralized Methods: Cons Optimal solution is computationally hard –Intractable for more than a few robots Makes unrealistic assumptions: –All relevant info can be transmitted to leader –This info doesn’t change during plan construction Vulnerable to malfunction of leader Heavy communication load for the leader

Market-Based Methods Based on market architecture Each robot seeks to maximize individual “profit” Robots can negotiate and bid for tasks Individual profit helps the common good Decisions are made locally but effects approach optimality –Preserves advantages of distributed approach

Why Is This Good? Robust to changing conditions –Not hierarchical –If a robot breaks, tasks can be re-bid to others Distributed nature allows for quick response Only local communication necessary Efficient resource utilization and role adoption Advantages of distributed system with optimality approaching centralized system

Architecture World is represented as a grid –Squares are unknown (0), occupied (+), or empty (-) Goals are squares in the grid for a robot to explore –Goal points to visit are the main commodity exchanged in market For any goal square in the grid: –Cost based on distance traveled to reach goal –Revenue based on information gained by reaching goal R = (# of unknown cells near goal) x (weighting factor) Team profit = sum of individual profits –When individual robots maximize profit, the whole team gains

Example World

Goal Selection Strategies Possible strategies: –Randomly select points, discard if already visited –Greedy exploration: Choose goal point in closest unexplored region –Space division by quadtree

Exploration Algorithm Algorithm for each robot: 1. Generate goals (based on goal selection strategy) 2. If OpExec (human operator) is reachable, check with OpExec to make sure goals are new to colony 3. Rank goals greedily based on expected profit 4. Try to auction off /bid goals to each reachable robot –If a bid is worth more than you would profit from reaching the goal yourself (plus a markup), sell it

Exploration Algorithm 5. Once all auctions are closed, explore highest-profit goal 6. Upon reaching goal, generate new goal points –Maximum # of goal points is limited 7. Repeat this algorithm until map is complete

Bidding Example R1 auctions goal to R2

Expected vs. Real Robots make decisions based on expected profit –Expected cost and revenue based on current map Actual profit may be different –Unforeseen obstacles may increase cost Once real costs exceed expected costs by some margin, abandon goal –Don’t get stuck trying for unreachable goals

Information Sharing If an auctioneer tries to auction a goal point already covered by a bidder: –Bidder tells auctioneer to update map –Removes goal point Robots can sell map information to each other –Price negotiated based on information gained –Reduces overlapping exploration When needed, OpExec sends a map request to all reachable robots –Robots respond by sending current maps –OpExec combines the maps by adding up cell values

Advantages of Communication Low-bandwidth mechanisms for communicating aggregate information Unlike other systems, map info doesn’t need to be communicated repeatedly for coordination

What Is a Robot Doing Goal generation and exploration Sharing Information with other robots Report information to OpExec at some frequency

Experimental Setup 4 or 5 robots –Equipped with fiber optic gyroscopes –16 ultrasonic sensors

Experimental Setup Three test environments –Large room cluttered with obstacles –Outdoor patio, with open areas as well as walls and tables –Large conference room with tables and 100 people wandering around Took between 5 and 10 minutes to map areas

Experimental Results

Successfully mapped regions Performance metric (exploration efficiency): –Area covered / distance traveled [m 2 / m] –Market architecture improved efficiency over no communication by a factor of 3.4

Conclusion Market-based approach for multi-robot coordination is promising –Robustness and quickness of distributed system –Approaches optimality of centralized system –Low communication requirements Probably not perfect –Cost heuristics can be inaccurate –Much of this approach is still speculative Some pieces, such as leaders, may be too hard to do

In Sum Distributed vs. centralized mapping Distributed vs. centralized planning Revenue/Cost vs. Revenue – Cost Often sub-optimal solutions No common evaluation system for comparisons

References Yamauchi, B., "Frontier-Based Exploration Using Multiple Robots," In Proc. of the Second International Conference on Autonomous Agents (Agents98), Minneapolis, MN., Menglei Jia, Guangming Zhou,Zonghai Chen, "Arena—an Architecture for Multi-Robot Exploration Combining Task Allocation and Path Planning,“ 2004 Zlot, R., Stentz, A., Dias, M. B., and Thayer, S. “ Multi-Robot Exploration Controlled By A Market Economy. ” Proceedings of the IEEE International Conference on Robotics and Automation, loration.ppt loration.ppt