Optimization Based Approaches to Autonomy March 3, 2005 Cedric Ma Northrop Grumman Corporation SAE Aerospace Control and Guidance Systems Committee (ACGSC)

Slides:



Advertisements
Similar presentations
Formal Computational Skills
Advertisements

Reactive and Potential Field Planners
ARCHITECTURES FOR ARTIFICIAL INTELLIGENCE SYSTEMS
March 13, 2006 © 2006 Kettle Creek Corporation 1 The OODA “Loop” Sketch Note how orientation shapes observation, shapes decision, shapes action, and in.
Lect.3 Modeling in The Time Domain Basil Hamed
Model Checker In-The-Loop Flavio Lerda, Edmund M. Clarke Computer Science Department Jim Kapinski, Bruce H. Krogh Electrical & Computer Engineering MURI.
A discussion on.  Path Planning  Autonomous Underwater Vehicles  Adaptive Sampling  Mixed Integer Linear programming.
Resource Management of Highly Configurable Tasks April 26, 2004 Jeffery P. HansenSourav Ghosh Raj RajkumarJohn P. Lehoczky Carnegie Mellon University.
Hierarchical mission control of automata with human supervision Prof. David A. Castañon Boston University.
Motion planning, control and obstacle avoidance D. Calisi.
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
Infinite Horizon Problems
Where in the World Are We? From the Voluntary Sector Round Table.
Robust Hybrid and Embedded Systems Design Jerry Ding, Jeremy Gillula, Haomiao Huang, Michael Vitus, and Claire Tomlin MURI Review Meeting Frameworks and.
Nonholonomic Multibody Mobile Robots: Controllability and Motion Planning in the Presence of Obstacles (1991) Jerome Barraquand Jean-Claude Latombe.
Multiple Marine Vehile Deconflicted Path Planning with Currents and Communication Constraints A. Häusler 1, R. Ghabcheloo 2, A. Pascoal 1, A. Aguiar 1.
Robotics CSPP Artificial Intelligence March 10, 2004.
Modeling and Planning with Robust Hybrid Automata Cooperative Control of Distributed Autonomous Vehicles in Adversarial Environments 2001 MURI: UCLA, CalTech,
P. Ögren (KTH) N. Leonard (Princeton University)
February 24, Final Presentation AAE Final Presentation Backstepping Based Flight Control Asif Hossain.
Sense & Avoid for UAV Systems
Efficient Methodologies for Reliability Based Design Optimization
Decentralized Optimization, with application to Multiple Aircraft Coordination Decision Making Under Uncertainty MURI Review, July 2002 Gökhan Inalhan,
Continuum Crowds Adrien Treuille, Siggraph 王上文.
Chapter 5: Path Planning Hadi Moradi. Motivation Need to choose a path for the end effector that avoids collisions and singularities Collisions are easy.
Multi-vehicle Cooperative Control Raffaello D’Andrea Mechanical & Aerospace Engineering Cornell University u Progress on RoboFlag Test-bed u MLD approach.
Aeronautics & Astronautics Autonomous Flight Systems Laboratory All slides and material copyright of University of Washington Autonomous Flight Systems.
Decentralised Coordination of Mobile Sensors School of Electronics and Computer Science University of Southampton Ruben Stranders,
Definition of an Industrial Robot
Constraints-based Motion Planning for an Automatic, Flexible Laser Scanning Robotized Platform Th. Borangiu, A. Dogar, A. Dumitrache University Politehnica.
World space = physical space, contains robots and obstacles Configuration = set of independent parameters that characterizes the position of every point.
Formulation of a complete structural uncertainty model for robust flutter prediction Brian Danowsky Staff Engineer, Research Systems Technology, Inc.,
© Manfred Huber Autonomous Robots Robot Path Planning.
A Framework for Distributed Model Predictive Control
August 2006 © 2006 Kettle Creek Corporation 1 The OODA “Loop” Sketch Note how orientation shapes observation, shapes decision, shapes action, and in turn.
1 [Click (left) mouse button to advance] The OODA “Loop” Sketch Note how orientation shapes observation, shapes decision, shapes action, and in turn is.
Robotics Chapter 5 – Path and Trajectory Planning
ASAS FRA OB/T ATM Projects Lufthansa point of view.
Engineering Systems Analysis for Design Richard de Neufville © Massachusetts Institute of Technology Dynamic Programming Slide 1 of 35 Dynamic Programming.
NUS CS5247 Deadlock-Free and Collision-Free Coordination of Two Robot Manipulators By Patrick A. O’Donnell and Tomás Lozano-Pérez MIT Artificial Intelligence.
MURI: Integrated Fusion, Performance Prediction, and Sensor Management for Automatic Target Exploitation 1 Dynamic Sensor Resource Management for ATE MURI.
Approximate Dynamic Programming Methods for Resource Constrained Sensor Management John W. Fisher III, Jason L. Williams and Alan S. Willsky MIT CSAIL.
1 Distributed and Optimal Motion Planning for Multiple Mobile Robots Yi Guo and Lynne Parker Center for Engineering Science Advanced Research Computer.
S ystems Analysis Laboratory Helsinki University of Technology Automated Solution of Realistic Near-Optimal Aircraft Trajectories Using Computational Optimal.
UAV See & Avoid Employing Vision Sensors
Issues and Challenges for Co-operative UAV Missions Chris Halliday and Tony Dodd.
March 2004 At A Glance autoProducts is an automated flight dynamics product generation system. It provides a mission flight operations team with the capability.
Outline: Introduction Solvability Manipulator subspace when n<6
Optimal Path Planning Using the Minimum-Time Criterion by James Bobrow Guha Jayachandran April 29, 2002.
Autonomous Robots Robot Path Planning (3) © Manfred Huber 2008.
September 28, 2000 Improved Simultaneous Data Reconciliation, Bias Detection and Identification Using Mixed Integer Optimization Methods Presented by:
Path Planning Based on Ant Colony Algorithm and Distributed Local Navigation for Multi-Robot Systems International Conference on Mechatronics and Automation.
Vision-Guided Humanoid Footstep Planning for Dynamic Environments P. Michel, J. Chestnutt, J. Kuffner, T. Kanade Carnegie Mellon University – Robotics.
1 Source Internet (Where else?) – Defense and the National Interest (a vetted source) For MANY additional aspects.
Beard & McLain, “Small Unmanned Aircraft,” Princeton University Press, 2012, Chapter 12: Slide 1 Chapter 12 Path Planning.
1 Chapter 5 Branch-and-bound Framework and Its Applications.
CSPP Artificial Intelligence March 10, 2004
Vision-Guided Humanoid Footstep Planning for Dynamic Environments
Optimal Acceleration and Braking Sequences for Vehicles in the Presence of Moving Obstacles Jeff Johnson, Kris Hauser School of Informatics and Computing.
CMSC Artificial Intelligence March 11, 2008
Friends and Partners of Aviation Weather
The Relationship Between Traffic Stability and Capacity for Decentralized Airspace 7th International Conference for Research in Air Transportation June.
C.-S. Shieh, EC, KUAS, Taiwan
Optimal UAV Flight Path Selection
Autonomous Cyber-Physical Systems: Dynamical Systems
Robust Belief-based Execution of Manipulation Programs
Real-Time Motion Planning
Self-Managed Systems: an Architectural Challenge
Chapter 4 . Trajectory planning and Inverse kinematics
Source Internet (Where else?) – Defense and the National Interest (a vetted source) For MANY additional aspects.
Presentation transcript:

Optimization Based Approaches to Autonomy March 3, 2005 Cedric Ma Northrop Grumman Corporation SAE Aerospace Control and Guidance Systems Committee (ACGSC) Meeting Harvey’s Resort, Lake Tahoe, Nevada

2 Outline  Introduction  Level of Autonomy  Optimization and Autonomy  Autonomy Hierarchy and Applications  Path Planning with Mixed Integer Linear Programming  Optimal Trajectory Generation with Nonlinear Programming  Summary and Conclusions

3 Autonomy in Vehicle Applications FORMATION FLYING PACK LEVEL COORDINATION RENDEZVOUS & REFUELING OBSTACLE AVOIDANCE COOPERATIVE SEARCH NAVIGATION TEAM TACTICS LANDING

4 Autonomy: Boyd’s OODA “Loop” Note how orientation shapes observation, shapes decision, shapes action, and in turn is shaped by the feedback and other phenomena coming into our sensing or observing window. Also note how the entire “loop” (not just orientation) is an ongoing many-sided implicit cross-referencing process of projection, empathy, correlation, and rejection. From “The Essence of Winning and Losing,” John R. Boyd, January Note how orientation shapes observation, shapes decision, shapes action, and in turn is shaped by the feedback and other phenomena coming into our sensing or observing window. Also note how the entire “loop” (not just orientation) is an ongoing many-sided implicit cross-referencing process of projection, empathy, correlation, and rejection. From “The Essence of Winning and Losing,” John R. Boyd, January Feed Forward Observations Decision (Hypothesis) Action (Test) Cultural Traditions Genetic Heritage New Information Previous Experience Analyses & Synthesis Feed Forward Implicit Guidance & Control Unfolding Interaction With Environment Feedback Outside Information Unfolding Circumstances ObserveOrientDecideAct Defense and the National Interest,

5 Level of Autonomy  Ground Operation  Activities performed off-line  Tele-Operation  Awareness of sensor / actuator interfaces  Executes commands uploaded from the ground  Reactive Control  Awareness of the present situation  Simple reflexes, i.e. no planning required  A condition triggers an associated action  Responsive Control  Awareness of past actions  Remembers previous actions  Remembers features of the environment  Remembers goals  Deliberative Control  Awareness of future possibilities  Reasons about future consequences  Chooses optimal paths / plans Ground operation Tele-operation Reactive Control Responsive Control Deliberative Control Goal of Optimization Based Autonomy

6 Optimal Control/Decision Objective/Reward Function Constraints/Rules (i.e. Dynamics/Goal) Optimization and Autonomy Optimizer Vehicle State Determines best course of action based on current objective, while meeting constraints Formulation of problem shapes the “Orient” mechanism

7 Autonomy Hierarchy Cooperative Control Mission Planning Path Planning Trajectory Generation Trajectory Following/ Inner Loop Planning & Scheduling, Resource Allocation & Sequencing Task Sequencing, Auto Routing Time Scale: ~1 hr Multi-Agent Coordination, Pack Level Organization Formation Flying, Cooperative Search & Electronic Warfare Conflict Resolution, Task Negotiation, Team Tactics Time Scale: ~1 min “Navigation,” Motion Planning Obstacle/Collision/Threat Avoidance Time Scale: ~10s “Guidance,” Contingency Handling Landing, Rendezvous, Refueling Time Scale: ~1s “Control,” Disturbance Rejection Applications: Stabilization, Adaptive Reconfigurable Control, FDIR Time Scale: ~0.1s

8 Path Planning with Mixed-Integer Linear Programming (MILP)

9 Overview: Path Planning  Path Planning bridges the gap between Mission Planner/AutoRouter and Individual Vehicle Guidance  Acts on an “intermediate” time scale between that of mission planner (minutes) and guidance (<seconds)  Short reaction time Mission waypoints Collision Avoidance Obstacle Avoidance Terrain Navigation Multi-vehicle Coordination Nap-of-the-Earth Flight

10 Path-Planning with MILP  Mixed-Integer Linear Programming  Linear Programs (LP) with integer variables  COTS MILP solver: ILOG CPLEX  Vehicle dynamics as linear constraints:  Limit velocity, acceleration, climb/turn rate  Resulting path is given to 4-D guidance  Integer variables can model:  Obstacle collision constraints (binary)  Control Modes, Threat Exposure  Nonlinear Functions: RCS, Dynamics  Min. Time, Acceleration, Altitude, Threat  Objective function includes terms for: Acceleration, Non-Arrival, Terminal, Altitude, Threat Exposure

11 Basic Obstacle Avoidance Problem  Vehicle Dynamic Constraints  Double Integrator dynamics  Max acceleration  Max velocity  Objective Function (summed over each time step)  Acceleration (1-norm) in x, y, z  Distance to destination (1-norm)  Altitude (if applicable)  Obstacle Constraints (integer)  One set per obstacle per time step  No cost associated with obstacles x – M b 1 ≤ x 1 x + M b 2 ≥ x 2 y – M b 3 ≤ y 1 y + M b 4 ≥ y 2 b 1 + b 2 + b 3 + b 4 ≤ 3

12 Receding Horizon MILP Path-Planning  Path is computed periodically, with most current information  Planning horizon, replan period chosen based on problem type, computational requirements, & environment  Only subset of current plan is executed before replanning  RH reduces computation time  Shorter planning horizon  Does not plan to destination  RH introduces robustness to path planning  Pop-up obstacles  Unexpected obstacle movement

13 Obstacle Avoidance Treetop level Nap of the Earth Flight Urban Low Altitude Operations

14 Collision Avoidance  Problem is formulated identically as Obstacle Avoidance in MILP  Air vehicles are moving obstacles  Path calculation based on expected future trajectory of other vehicles  Dealing with Uncertainty  Vehicles of uncertain intent can be enlarged with time  Receding Horizon  Frequent replanning  Change in planned path (blue) in response to changes in intruder movement

15 Coordinated Conflict Resolution  3-D Multi-Vehicle Path- Planning problem  Centralized version  “Decentralized Cooperative Trajectory Planning of Multiple Aircraft with Hard Safety Guarantees” by MIT  Loiter maneuvers can be used to produce provably safe trajectories  Minimum separation distance is specified in problem formulation  No limit to number of vehicles  Non-cooperative vehicles are treated as moving obstacles

16 Threat Avoidance  Purpose: To avoid detection by known threats by planning trajectory behind opaque obstacles  Shadow-like “Safe Zones”  One per threat/obstacle pair  Well defined for convex obstacles  Nice topological properties  Patent Pending: Docket No Threat Vehicle hiding behind building On-time arrival at destination

17 MILP Path Planning  MILP: Fast Global Optimization  No suboptimal local minima  Branch & Bound provides fast tree-search  Commercial solver on RTOS  Tractability Trade-off:  Time Discretization  Constraints active only at discrete points in time  Time Scale Refinement  Linear dynamics/constraints  Formulation should properly capture nonlinearity of solution space  True global minimum is in a neighborhood of MILP optimal solution Summary

18 Optimal Trajectory Generation with Nonlinear Programming (NLP)

19 Problems & Goal of Trajectory Generation  Currently, the primary method is pre-generated waypoint routes with little/no adaptation or reaction to threats or condition changes  Even the latest vehicles have low autonomy levels and are doing exactly what they are told, largely indifferent to the world around them  What are the potential gains of Near Real Time Trajectory Generation?  Improved Effectiveness  Reduced operator workload – force multiplier  Mission planning / re-planning  Account for range and time delays  Improved Survivability?  UAV trades success/risk  Limp-home capability  Autonomous threat mitigation (RCS, SAM, Small Arms, AA Fire)  Air/Air Engagement  Accurate release of cheap ‘dumb’ ordinance  GOAL DRIVEN AUTONOMY Command ‘What’ not ‘How’  How best can we mimic (improve?) on human skill and speed at trajectory generation in complex environments?

20 Classical Trajectory Optimization Problem Issues: Becomes the traditional two point constrained boundary value problem Computationally expensive due to equality constraints from the system, environment and actuation dynamics Currently intractable in required time for effective control Hope? Perhaps our systems contain a structure which allows all solutions of the system, (trajectories) to be smoothly mapped, from a set of free trajectories in a reduced dimensional space. Algebraic solutions in this reduced space would implicitly satisfying the dynamic constraints of the original system. Cost: Constraints:

21 Trajectory Generation: Current Methods  Brute force numerical method solution of the dynamic and constraint ODE’s  Solution Method 1) Guess control  e (t) 2) Propagate dynamics from beginning to end (simulate) 3) Propagate constraints from beginning to end (simulate) 4) Check for constraint violation 5) Modify guess  e (t) 6) Repeat until feasible/optimal solution obtained. (optimize)  Vast complexity and extremely long solution times are addressed by either/both:  Very simple control curves  All calculations performed offline (selected/looked-up online)  Much of previous work in subject devoted to improving ‘wisdom’ of next guess  ee  ee Iteration 1: Iteration 2:

22 Differential Systems Suggest an Elegant Solution  Perhaps our systems contain a structure which allows all solutions of the system, (trajectories) to be smoothly mapped by a set of free trajectories in a reduced dimensional space. Algebraic solutions in this reduced space would implicitly satisfying the dynamic constraints of the original system dynamics and constraint ODE’s  Constraints are mapped into the flat space as well and also become time independent  Direct Solutions! We are modifying the same curve we are optimizing!  Local Support: Every solution is only affected by the trajectory near it  Basically a curve fit problem

23 Definition: A system is said to be differentially flat if there exists variables z 1,…,z m of the form such that (x,u) can be expressed in terms of z and its derivatives by an equation of the form Note: Dynamic Feedback Linearization via endogenous feedback is equivalent to differential flatness. Example: (Point-to-Point): Differential Constraints are reduced to algebraic equations in the Flat space! Differential Flatness

24 Instinct Autonomy: Now Using Flatness  Simply find any curve that satisfies the constraints in the flat space  Solution Method 1) Map system to flat space using ‘w -1 ’ 2) Guess trajectory of flat output z n 3) Compare against constraints (in flat space) 4) Optimize over control points  When completed apply ‘w’ function to convert back to normal space  Much simpler control space, no simulation required:  Very simple to manipulate curves  All calculations performed on- line on the vehicle z2z2 ee  z1z1

25 Too Good to be True? What did we Lose?  It seems reasonable that such a reduction in complexity would result in some sort of approximation  Many systems lose nothing at all!  Linear models that are controllable (including non- minimum phase)  Fully-flat nonlinear models  Some systems make reasonable assumptions  Conventional A/C make identical assumptions as dynamic inversion  Some systems are very much less obvious and more complicated  This is one of the hardest questions of Differential Flatness – identifying the flat output can be very difficult  Modern configurations are very challenging!  After one stabilization loop, most systems become differentially flat (or very close to it)

26 SEC Autonomous Trajectory Generation GO TO WP_C GO TO Rnwy_3 GO TO WP_A GO TO WP_D GO TO WP_S

27 MILP Path Planning  MILP: Fast Global Optimization  No suboptimal local minima  Branch & Bound provides fast tree-search  Commercial solver on RTOS  Tractability Trade-off:  Time Discretization  Constraints active only at discrete points in time  Time Scale Refinement  Linear dynamics/constraints  Formulation should properly capture nonlinearity of solution space  True global minimum is in a neighborhood of MILP optimal solution Optimal Trajectory Generation  OTG: Fast Nonlinear Optimization  Optimal control for full nonlinear systems  Differential Flatness property allows problem to be mapped to lower dimensional space for NLP solver  Absence of dynamics in new space speeds optimization  Easier constraint propagation  Problem setup should focus on right “basin of attraction”  NLP solver seeks locally optimal solutions via SQP methods  Good initial guess  Use in conjunction with global methods, i.e. MILP Summary

28 Conclusions  Optimization based approaches help achieve a higher level of autonomy by enabling autonomous decision making  Cast autonomy applications into standard optimization problems, to be solved using existing optimization tools and framework  Benefits: no need to build custom solver, existing body of theory, continued improvement in solver technology  Future: broad range of complex autonomy applications are enabled by a wide, continuous spectrum of powerful optimization engines and approaches  Challenge: advanced development of V&V, sensing, & fusion technology, leading to widespread certification and adoption  Thanks/Credits:  NTG/OTG Approach: Mark Milam/NGST, Prof. R. Murray/Caltech  MILP Approach: Prof. Jonathan How/MIT  Autonomy Slides: Jonathan Mead/NGST  OTG Slides: Travis Vetter/NGIS