Analytics and OR DP- summary.

Slides:



Advertisements
Similar presentations
Markov Decision Process
Advertisements

SA-1 Probabilistic Robotics Planning and Control: Partially Observable Markov Decision Processes.
CSE-573 Artificial Intelligence Partially-Observable MDPS (POMDPs)
CompLACS Composing Learning for Artificial Cognitive Systems Year 2: Specification of scenarios.
Planning under Uncertainty
1 Solving problems by searching Chapter 3. 2 Why Search? To achieve goals or to maximize our utility we need to predict what the result of our actions.
Math443/543 Mathematical Modeling and Optimization
Nov 14 th  Homework 4 due  Project 4 due 11/26.
Models Physical: Scale, Analog Symbolic: Drawings Computer Programs Mathematical: Analytical (Deduction) Experimental (Induction)
Strategic Decisions Using Dynamic Programming
Simulation Models as a Research Method Professor Alexander Settles.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Optimal Fixed-Size Controllers for Decentralized POMDPs Christopher Amato Daniel.
Chapter 01 Introduction to Probability Models Course Focus Textbook Approach Why Study This?
Robert M. Saltzman © DS 851: 4 Main Components 1.Applications The more you see, the better 2.Probability & Statistics Computer does most of the work.
On the convergence of SDDP and related algorithms Speaker: Ziming Guan Supervisor: A. B. Philpott Sponsor: Fonterra New Zealand.
Utility Theory & MDPs Tamara Berg CS Artificial Intelligence Many slides throughout the course adapted from Svetlana Lazebnik, Dan Klein, Stuart.
MAKING COMPLEX DEClSlONS
Planning and Verification for Stochastic Processes with Asynchronous Events Håkan L. S. Younes Carnegie Mellon University.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
CSE-473 Artificial Intelligence Partially-Observable MDPS (POMDPs)
Approximation Algorithms Pages ADVANCED TOPICS IN COMPLEXITY THEORY.
An Overview of Dynamic Programming Seminar Series Joe Hartman ISE October 14, 2004.
1 Definition of System Simulation: The practice of building models to represent existing real-world systems, or hypothetical future systems, and of experimenting.
Decision Making in Robots and Autonomous Agents Decision Making in Robots and Autonomous Agents The Markov Decision Process (MDP) model Subramanian Ramamoorthy.
Managerial Decision Making and Problem Solving
Constraint Satisfaction Problems (CSPs) CPSC 322 – CSP 1 Poole & Mackworth textbook: Sections § Lecturer: Alan Mackworth September 28, 2012.
ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 31 Alternative System Description If all w k are given initially as Then,
1 ECE-517 Reinforcement Learning in Artificial Intelligence Lecture 7: Finite Horizon MDPs, Dynamic Programming Dr. Itamar Arel College of Engineering.
Copyright © 2012, SAS Institute Inc. All rights reserved. ANALYTICS IN BIG DATA ERA ANALYTICS TECHNOLOGY AND ARCHITECTURE TO MANAGE VELOCITY AND VARIETY,
© 2015 McGraw-Hill Education. All rights reserved. Chapter 19 Markov Decision Processes.
Lecture 3: 18/4/1435 Searching for solutions. Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
MDPs (cont) & Reinforcement Learning
Dynamic Programming Discrete time frame Multi-stage decision problem Solves backwards.
© 2009 Ilya O. Ryzhov 1 © 2008 Warren B. Powell 1. Optimal Learning On A Graph INFORMS Annual Meeting October 11, 2009 Ilya O. Ryzhov Warren Powell Princeton.
Decision Theoretic Planning. Decisions Under Uncertainty  Some areas of AI (e.g., planning) focus on decision making in domains where the environment.
1 Inventory Control with Time-Varying Demand. 2  Week 1Introduction to Production Planning and Inventory Control  Week 2Inventory Control – Deterministic.
1 1 Slide Simulation Professor Ahmadi. 2 2 Slide Simulation Chapter Outline n Computer Simulation n Simulation Modeling n Random Variables and Pseudo-Random.
Reinforcement Learning Dynamic Programming I Subramanian Ramamoorthy School of Informatics 31 January, 2012.
DEPARTMENT/SEMESTER ME VII Sem COURSE NAME Operation Research Manav Rachna College of Engg.
Csci 418/618 Simulation Models Dr. Ken Nygard, IACC 262B
Planning Under Uncertainty. Sensing error Partial observability Unpredictable dynamics Other agents.
Generalized Point Based Value Iteration for Interactive POMDPs Prashant Doshi Dept. of Computer Science and AI Institute University of Georgia
Definition of the Hidden Markov Model A Seminar Speech Recognition presentation A Seminar Speech Recognition presentation October 24 th 2002 Pieter Bas.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Achieving Goals in Decentralized POMDPs Christopher Amato Shlomo Zilberstein UMass.
Modelling & Simulation of Semiconductor Devices Lecture 1 & 2 Introduction to Modelling & Simulation.
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 3
Prof. Dr. Holger Schlingloff 1,2 Dr. Esteban Pavese 1
CS b659: Intelligent Robotics
Making complex decisions
POMDPs Logistics Outline No class Wed
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 3
Policy Gradient in Continuous Time
Optimization Techniques for Natural Resources SEFS 540 / ESRM 490 B
Clearing the Jungle of Stochastic Optimization
Markov Decision Processes
Jianping Fan Dept of CS UNC-Charlotte
Markov Decision Processes
Course Logistics CS533: Intelligent Agents and Decision Making
Announcements Homework 3 due today (grace period through Friday)
INNOvation in TRAINING BUSINESS ANALYSTS HAO HElEN Zhang UniVERSITY of ARIZONA
Markov Decision Problems
MECH 3550 : Simulation & Visualization
Russell and Norvig: Chapter 3, Sections 3.1 – 3.3
Introduction to Modeling
CS 416 Artificial Intelligence
Reinforcement Learning Dealing with Partial Observability
CS 416 Artificial Intelligence
Area Coverage Problem Optimization by (local) Search
Dr. Arslan Ornek MATHEMATICAL MODELS
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 3
Presentation transcript:

Analytics and OR DP- summary

Prescriptive (Application of OR methods) Analytics Descriptive (IT, CS) Prescriptive (Application of OR methods) Predictive (STAT) Theory and methods of OR OR

Thinking like an OR Analyst Linear or non-linear Dynamic or Static Probabilistic or Deterministic One time decision or sequential decisions Objective What is the objective of making the decision(s) (Value Modeling can be used here) Big Data Availability Type (Static, dynamic (frequency of collection), discrete, continuous) Size (Big Data) Model Known, unknown (Learning) Which OR tool is appropriate Solution Strategy Computational complexity Verification Validation Implementation Decision making in optimization Problem Definition

The Big Picture Decisions taken over time as the system evolves Operations Research (Math prog) Static decisions (One-time decision) Linear Programming Integer Non-linear Mixed Integer Metaheuristics Dynamic decisions (sequential Dynamic programming Deterministic DP Finite Horizon problems Infinite Horizon Stochastic The Big Picture Decisions taken over time as the system evolves Optimal control Continuous time ODE Approach Discrete time Large-scale problems Approximate Dynamic programming

Dynamic programming What is it? Operates by finding the shortest path (minimize cost) or longest path (maximize reward) of decision making problems that are solved over time

Myopic vs DP thinking - find shortest path from A to B 1 Cij=10 40 Cij= cost on the arc A B 2 10 20 Myopic policy: V(A)= min (Cij) = min of (10 or 20) leads to solution of 50 from A to 1 to B DP policy: V(A) = min (Cij + V(next node)) = min (10 + 40, 20+10) = 30 leads to solution of 30 from A to 2 to B Key is to find the values of node 1 and 2 How? By learning via simulation-based optimization

Sequential Decision Making (over time) Dynamic Programming!! for Sequential Decision Making (over time) Is based on the idea that We want to move from one good state of the system to another by making an near-optimal decision in the presence of uncertainty It also means that the future state depends only on the current and the decision taken in the presence of uncertainly (and not on the past – memory-less property) The above is achieved via unsupervised learning that entails only an interaction with the environment in a model-free setting

Simulation-based Optimization In more complex problems such as helicopter control the Optimizer is an Artificial Intelligence (Learning) Agent Environment (uncertainty) Output (Objective function) System simulator built using probability distributions from real-world data Decisions Dynamic Programming Optimizer (learning agent)

In Summary OR OR, CS IT OR STAT IT, CS Analytics Descriptive (IT, CS) Prescriptive (Application of OR methods) Predictive (STAT) Theory and methods of OR OR Optimization in Prescriptive Analytics OR Models Computational complexity Algorithms for solving BIG Data and data mining Data Visualization & Statistical analysis Data storage OR, CS IT OR STAT IT, CS

In Summary – Main take away In Big Data decision making Problems (Prescriptive Analytics) Understand characteristics of the data, linear/non-linear, deterministic or probabilistic, static or dynamic (frequency of collection) Beware of 1. Myopic policies 2. Exhaustive enumeration of all solutions 3. Computational Complexity Look for appropriate modeling and solution strategies that can provide near-optimal decisions (good-enough) In the long run for the problem at hand.

For this course - Main take away Value function V is cumulative When making a decision sequentially over time (dynamic programming), make sure to sum the cost/reward of making the decision with the value of the estimated future state that the decision will take you to. Then pick a decision that minimizes (if cost) or maximizes (if reward) the above sum. DP can handle -non-linear -probabilistic -Computationally hard (high dimensionality and large state space) -dynamic (sequential) and high frequency of decision making -unknown models -Big data problems In this course we solve optimally In OR 774 and in real world we strive of near-optimal (good enough) solutions

Thank You