Zach Ramaekers Computer Science University of Nebraska at Omaha Advisor: Dr. Raj Dasgupta 1.

Slides:



Advertisements
Similar presentations
IMA 2.5: Software Architecture and Development Environment Roberto Olivares M.S. Electrical Engineering Vanderbilt University, Spring 2003.
Advertisements

1 Toward a Modeling Theory for Predictable Complex Software Designs by Levent Yilmaz Auburn Modeling and Simulation Laboratory Department of Computer Science.
Motion Planning for Point Robots CS 659 Kris Hauser.
Weighted Voting Game Based Multi-robot Team Formation for Distributed Area Coverage Ke Cheng and Prithviraj (Raj) Dasgupta Computer Science Department.
SA-1 Probabilistic Robotics Planning and Control: Partially Observable Markov Decision Processes.
Carl Nelson*, Khoa Chu*, Prithviraj (Raj) Dasgupta**, Zachary Ramaekers** University of Nebraska *: Mechanical Engineering, University of Nebraska, Lincoln.
Dealing with Complexity Robert Love, Venkat Jayaraman July 24, 2008 SSTP Seminar – Lecture 10.
Optimal Policies for POMDP Presented by Alp Sardağ.
COORDINATION and NETWORKING of GROUPS OF MOBILE AUTONOMOUS AGENTS.
4/15/2017 Using Gaussian Process Regression for Efficient Motion Planning in Environments with Deformable Objects Barbara Frank, Cyrill Stachniss, Nichola.
Planning under Uncertainty
1 Sensor Networks and Networked Societies of Artifacts Jose Rolim University of Geneva.
AAMAS 2009, Budapest1 Analyzing the Performance of Randomized Information Sharing Prasanna Velagapudi, Katia Sycara and Paul Scerri Robotics Institute,
Zachary Wilson Computer Science Department University of Nebraska, Omaha Advisor: Dr. Raj Dasgupta.
Adaptive Multi-Robot Team Reconfiguration using a Policy-Reuse Reinforcement Learning Approach Ke Cheng 1, Raj Dasgupta 1 and Bikramjit Banerjee 2 1 Computer.
Decentralised Coordination of Mobile Sensors using the Max-Sum Algorithm Ruben Stranders, Alex Rogers, Nick Jennings School of Electronics and Computer.
Localized Techniques for Power Minimization and Information Gathering in Sensor Networks EE249 Final Presentation David Tong Nguyen Abhijit Davare Mentor:
A Principled Information Valuation for Communications During Multi-Agent Coordination Simon A. Williamson, Enrico H. Gerding, Nicholas R. Jennings School.
Integrating POMDP and RL for a Two Layer Simulated Robot Architecture Presented by Alp Sardağ.
Simulation Models as a Research Method Professor Alexander Settles.
Multiagent Planning with Factored MDPs Carlos Guestrin Daphne Koller Stanford University Ronald Parr Duke University.
A Game-Theoretic Approach to Determining Efficient Patrolling Strategies for Mobile Robots Francesco Amigoni, Nicola Gatti, Antonio Ippedico.
Exploration in Reinforcement Learning Jeremy Wyatt Intelligent Robotics Lab School of Computer Science University of Birmingham, UK
Mobile Robot ApplicationsMobile Robot Applications Textbook: –T. Bräunl Embedded Robotics, Springer 2003 Recommended Reading: 1. J. Jones, A. Flynn: Mobile.
Tal Saiag & Anna Itin May 2013
ModRED: A Modular Self-Reconfigurable Robot for Autonomous Exploration Carl Nelson*, Khoa Chu*, Prithviraj (Raj) Dasgupta** University of Nebraska *: Mechanical.
Generating Adaptation Policies for Multi-Tier Applications in Consolidated Server Environments College of Computing Georgia Institute of Technology Gueyoung.
COGNITIVE RADIO FOR NEXT-GENERATION WIRELESS NETWORKS: AN APPROACH TO OPPORTUNISTIC CHANNEL SELECTION IN IEEE BASED WIRELESS MESH Dusit Niyato,
Artificial Intelligence Dr. Paul Wagner Department of Computer Science University of Wisconsin – Eau Claire.
Chapter 14: Artificial Intelligence Invitation to Computer Science, C++ Version, Third Edition.
Cooperating AmigoBots Framework and Algorithms
Leslie Luyt Supervisor: Dr. Karen Bradshaw 2 November 2009.
Stochastic Algorithms Some of the fastest known algorithms for certain tasks rely on chance Stochastic/Randomized Algorithms Two common variations – Monte.
1 ECE-517 Reinforcement Learning in Artificial Intelligence Lecture 7: Finite Horizon MDPs, Dynamic Programming Dr. Itamar Arel College of Engineering.
TKK | Automation Technology Laboratory Partially Observable Markov Decision Process (Chapter 15 & 16) José Luis Peralta.
Reinforcement Learning Presentation Markov Games as a Framework for Multi-agent Reinforcement Learning Mike L. Littman Jinzhong Niu March 30, 2004.
Dynamic Programming for Partially Observable Stochastic Games Daniel S. Bernstein University of Massachusetts Amherst in collaboration with Christopher.
Page 1 Reconfigurable Communications Processor Principal Investigator: Chris Papachristou Task Number: NAG Electrical Engineering & Computer Science.
Boundary Assertion in Behavior-Based Robotics Stephen Cohorn - Dept. of Math, Physics & Engineering, Tarleton State University Mentor: Dr. Mircea Agapie.
1 Distributed and Optimal Motion Planning for Multiple Mobile Robots Yi Guo and Lynne Parker Center for Engineering Science Advanced Research Computer.
MAPLD 2005/254C. Papachristou 1 Reconfigurable and Evolvable Hardware Fabric Chris Papachristou, Frank Wolff Robert Ewing Electrical Engineering & Computer.
Complexity of Determining Nonemptiness of the Core Vincent Conitzer, Tuomas Sandholm Computer Science Department Carnegie Mellon University.
Real-Time Support for Mobile Robotics K. Ramamritham (+ Li Huan, Prashant Shenoy, Rod Grupen)
Adaptive Hopfield Network Gürsel Serpen Dr. Gürsel Serpen Associate Professor Electrical Engineering and Computer Science Department University of Toledo.
UNC Chapel Hill M. C. Lin Introduction to Motion Planning Applications Overview of the Problem Basics – Planning for Point Robot –Visibility Graphs –Roadmap.
MDPs (cont) & Reinforcement Learning
KAIS T On the problem of placing Mobility Anchor Points in Wireless Mesh Networks Lei Wu & Bjorn Lanfeldt, Wireless Mesh Community Networks Workshop, 2006.
Global Hybrid Control and Cooperative Mobile Robotics Yi Guo Center for Engineering Science Advanced Research Computer Science and Mathematics Division.
1 Random Disambiguation Paths Al Aksakalli In Collaboration with Carey Priebe & Donniell Fishkind Department of Applied Mathematics and Statistics Johns.
Generalized Point Based Value Iteration for Interactive POMDPs Prashant Doshi Dept. of Computer Science and AI Institute University of Georgia
Abstract LSPI (Least-Squares Policy Iteration) works well in value function approximation Gaussian kernel is a popular choice as a basis function but can.
MODULAR SELF -RECONFIGURABLE MULTI-FUNCTIONAL ROBOTIC SYSTEM.
Multi-robot
COGNITIVE APPROACH TO ROBOT SPATIAL MAPPING
Complexity of Determining Nonemptiness of the Core
基于多核加速计算平台的深度神经网络 分割与重训练技术
Distributed Algorithms for Mobile Sensor Networks
Analytics and OR DP- summary.
Ayan Dutta1, Prithviraj Dasgupta1, Carl Nelson2
Multi-Agent Exploration
Towards Next Generation Panel at SAINT 2002
Announcements Homework 3 due today (grace period through Friday)
Dr. Unnikrishnan P.C. Professor, EEE
Reinforcement Learning
Market-based Dynamic Task Allocation in Mobile Surveillance Systems
Hidden Markov Models (cont.) Markov Decision Processes
CS 416 Artificial Intelligence
Self-Managed Systems: an Architectural Challenge
Reinforcement Learning Dealing with Partial Observability
Area Coverage Problem Optimization by (local) Search
Presentation transcript:

Zach Ramaekers Computer Science University of Nebraska at Omaha Advisor: Dr. Raj Dasgupta 1

Modular Self-Reconfigurable Robots (MSRs) – What and Why An MSR is a type of robot that is composed of identical modules The modules connect together to form larger robots capable of performing complex tasks Why MSRs? Inexpensive and Simple Highly Adaptable Three main types of MSRs: Chain, Lattice, Hybrid 2

ModRED (Modular Robot for Exploration and Discovery) Novel 4 degrees of freedom design Gives improved dexterity Allows maneuver itself and get out of tight spaces’ ModRED sensors and actuators Arduino processor (for doing computations) IR sensors(for sensing how far obstacles are) Compass (which direction am I heading) Tilt sensor (what is my inclination) XBee radio (for wireless comm.) Designed by Dr. Nelson’s group, Mechanical Engineering, UNL 3 CAD diagram of robot Simulated robot in Webots

ModRED Movements in Fixed Configuration 4 All these movements are in a fixed configuration

Problem Addressed: Dynamic Self- Reconfiguration by MSRs 5 Why does an MSR need to reconfigure dynamically? Problem Statement: How can an MSR that needs reconfigure (e.g., after encountering an obstacle) determine 1. which other modules to combine with, and 2. the best configuration to form with those modules?... in an autonomous manner.

Coalition Games for Dynamic MSR Reconfiguration We propose a novel, coalition game theory based approach to address the problem of MSR self-reconfiguration A coalition game gives a set of rules that determine what teams a group of humans (or robots) will form between themselves Teams are guranteed to be stable: once teams are formed no one will want to change teams 6 Game Theoretic Layer (Coalition Game) Controller Layer (Gait-tables) Mediator Dynamic self- reconfiguration Movement in Fixed Configuration

Coalition Games for Dynamic MSR Reconfiguration We propose a novel, coalition game theory based approach to address the problem of MSR self-reconfiguration A coalition game gives a set of rules that determine what teams a group of humans (or robots) will form between themselves Teams are guranteed to be stable: once teams are formed no one will want to change teams 7 Game Theoretic Layer (Coalition Game) Controller Layer (Gait-tables) Mediator Dynamic self- reconfiguration Movement in Fixed Configuration For our scenario: Each module of an MSR is provided with embedded software called an agent that does the coalition game related calculations

Coalition Games for Dynamic MSR Reconfiguration We propose a novel, coalition game theory based approach to address the problem of MSR self-reconfiguration A coalition game gives a set of rules that determine what teams a group of humans (or robots) will form between themselves Teams are guranteed to be stable: once teams are formed no one will want to change teams 8 Game Theoretic Layer (Coalition Game) Controller Layer (Gait-tables) Mediator Dynamic self- reconfiguration Movement in Fixed Configuration For our scenario: Each module of an MSR is provided with embedded software called an agent that does the coalition game related calculations Our problem: How can we determine these teams or partitions or coalitions for our MSR problem?

Determining the Partitions Enumerate all possible partitions that includes all the agents - called coalition structures 9 {1} {2}{3}{4}{1, 2}{3}{4}{1} {2,3}{4} Example coalition structures with 4 agents  Each coalition structure is associated with a value  V(CS i ) = sum of utilities of each coalition in CS i  All the possible coalition structures are represented as a coalition structure graph V(CS i ) = u(1,2) + u(3) + u(4) V(CS i ) = u(1) + u(2,3) + u(4) V(CS i ) = u(1) + u(2) + u(3) + u(4)

Coalition Structure Graph (CSG) 10 Possible partitions of 4 modules (agents) Problem: Find the node (coaliltion structure) that has the highest value in this graph

Coalition Structure Graph (CSG) Finding the optimal coalition structure node in this graph is not easy! CSG has  (n n ) nodes, the search problem is hard (NP-complete) Approximation algorithm used to find the optimal CSG node (Sandholm 1999, Rahwan 2007, etc.): in exponential time. 11 Possible partitions of 4 modules (agents) Problem: Find the node (coaliltion structure) that has the highest value in this graph

Dynamic Reconfiguration under Uncertainty 12 I need to form another configuration We are here And we are here Which modules should I join with? Ridge in planned path

Modeling Uncertainty in Coalition Formation 13 {1} {2}{3}{4}{1, 2}{3}{4}{1} {2,3}{4} Possible Coalition Structures Conventional CSG V(CS i ) = u(1,2) + u(3) + u(4) (additive reward value) {1, 2}{3}{4} CSG with uncertainty V(CS i ) < u(1,2) + u(3) + u(4) (subadditive) V(CS i ) = u(1,2) + u(3) + u(4) (additive) V(CS i ) > u(1,2) + u(3) + u(4) (superadditive) {1, 2}{3}{4}

Modeling Uncertainty in Coalition Formation Dealing with uncertainty: Markov Decision Process (MDP) provide a mathematical model for robots or agents to determine their actions in the presence of uncertainty 14 {1} {2}{3}{4}{1, 2}{3}{4}{1} {2,3}{4} Possible Coalition Structures Conventional CSG V(CS i ) = u(1,2) + u(3) + u(4) (additive reward value) {1, 2}{3}{4} CSG with uncertainty V(CS i ) < u(1,2) + u(3) + u(4) (subadditive) V(CS i ) = u(1,2) + u(3) + u(4) (additive) V(CS i ) > u(1,2) + u(3) + u(4) (superadditive) {1, 2}{3}{4}

MDP-Based CSG Solution 15 Conventional CSG without uncertainty Modified CSG with uncertainty

Algorithm to find optimal coalition structure Pruning – used to reduce the number of nodes that are searched by the algorithm in the coalition structure graph Three strategies explored: Keep the optimal and two least optimal children; keep the optimal, median, and least optimal children; keep three random children 16 Set of modules information Generate Coalition Utility Values Generate Coalition Structure Graph Run Value Iteration and Determine Policy Run MDP Traversal to Find Optimal CS Optimal Coalition Structure

Simulation Results 17

Conclusions and Future Work Developed coalition game theory based algorithm for MSR self-reconfiguration Validated to work on accurate model of MSR called ModRED within Webots robotic simulator To the best of our knowledge First application of game theory to MSR self- reconfiguration problem First attempt at combining planning under uncertainty (MDP) with coalition games* Future work Investigate distributed models of planning under uncertainty (MMDPs, DEC-MDPs, etc.) Simulation of exploration and coverage on realistic terrains Integrate with hardware of ModRED robot 18 *: Suijs et al.(1999) defined a framework called stochastic coalition game, but using a different model involving agent types, which wasn’t validated empirically.

ModRED Self-Reconfiguration Simulation Demo 19