Math 6330: Statistical Consulting Class 8

Slides:



Advertisements
Similar presentations
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Advertisements

1 University of Southern California Keep the Adversary Guessing: Agent Security by Policy Randomization Praveen Paruchuri University of Southern California.
Using Parallel Genetic Algorithm in a Predictive Job Scheduling
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
THE HONG KONG UNIVERSITY OF SCIENCE & TECHNOLOGY CSIT 5220: Reasoning and Decision under Uncertainty L09: Graphical Models for Decision Problems Nevin.
Dynamic Bayesian Networks (DBNs)
Modeling and Analysis of Random Walk Search Algorithms in P2P Networks Nabhendra Bisnik, Alhussein Abouzeid ECSE, Rensselaer Polytechnic Institute.
Planning under Uncertainty
A: A Unified Brand-name-Free Introduction to Planning Subbarao Kambhampati Environment What action next? The $$$$$$ Question.
CS 589 Information Risk Management 23 January 2007.
1 Hybrid Agent-Based Modeling: Architectures,Analyses and Applications (Stage One) Li, Hailin.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
End of Chapter 8 Neil Weisenfeld March 28, 2005.
Ensemble Learning (2), Tree and Forest
RESEARCH A systematic quest for undiscovered truth A way of thinking
MAKING COMPLEX DEClSlONS
Search and Planning for Inference and Learning in Computer Vision
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
MBA7025_01.ppt/Jan 13, 2015/Page 1 Georgia State University - Confidential MBA 7025 Statistical Business Analysis Introduction - Why Business Analysis.
MBA7020_01.ppt/June 13, 2005/Page 1 Georgia State University - Confidential MBA 7020 Business Analysis Foundations Introduction - Why Business Analysis.
Module networks Sushmita Roy BMI/CS 576 Nov 18 th & 20th, 2014.
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
MDPs (cont) & Reinforcement Learning
1 1 Slide Simulation Professor Ahmadi. 2 2 Slide Simulation Chapter Outline n Computer Simulation n Simulation Modeling n Random Variables and Pseudo-Random.
1 Optimizing Decisions over the Long-term in the Presence of Uncertain Response Edward Kambour.
1 Chapter 17 2 nd Part Making Complex Decisions --- Decision-theoretic Agent Design Xin Lu 11/04/2002.
Competition II: Springleaf Sha Li (Team leader) Xiaoyan Chong, Minglu Ma, Yue Wang CAMCOS Fall 2015 San Jose State University.
Operations Research Models and Methods Advanced Operations Research Reference: Operations Research Models and Methods, Operations Research Models and Methods,
1 Automated Planning and Decision Making 2007 Automated Planning and Decision Making Prof. Ronen Brafman Various Subjects.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Math 6330: Statistical Consulting Class 9
Decision Analysis Lecture 4
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Machine Learning: Ensemble Methods
Math 6330: Statistical Consulting Class 7
Decision Analysis Lecture 7
Keep the Adversary Guessing: Agent Security by Policy Randomization
Math 7 Statistics and Probability Unit 4
Math 6330: Statistical Consulting Class 2
Math 6330: Statistical Consulting Class 6
Intelligent Agents (Ch. 2)
Who am I? Work in Probabilistic Machine Learning Like to teach 
Math 6330: Statistical Consulting Class 11
Math 6330: Statistical Consulting Class 5
Stochastic tree search and stochastic games
Nevin L. Zhang Room 3504, phone: ,
Online Multiscale Dynamic Topic Models
Business Modeling Lecturer: Ing. Martina Hanová, PhD.
CS Fall 2015 (Shavlik©), Midterm Topics
Boosting and Additive Trees (2)
Mastering the game of Go with deep neural network and tree search
Analytics and OR DP- summary.
Reinforcement Learning (1)
Supplement: Decision Making
Games with Chance Other Search Algorithms
Professor S K Dubey,VSM Amity School of Business
Clearing the Jungle of Stochastic Optimization
Announcements Homework 3 due today (grace period through Friday)
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
CASE − Cognitive Agents for Social Environments
Multi-Objective Optimization
Dr. Unnikrishnan P.C. Professor, EEE
13. Acting under Uncertainty Wolfram Burgard and Bernhard Nebel
Course Lab Introduction to IBM Watson Analytics
October 6, 2011 Dr. Itamar Arel College of Engineering
Market-based Dynamic Task Allocation in Mobile Surveillance Systems
Ensemble learning Reminder - Bagging of Trees Random Forest
Lecture 3: Environs and Algorithms
Model generalization Brief summary of methods
Dr. Arslan Ornek MATHEMATICAL MODELS
Presentation transcript:

Math 6330: Statistical Consulting Class 8 Tony Cox tcoxdenver@aol.com University of Colorado at Denver Course web site: http://cox-associates.com/6330/

Agenda Projects and schedule Prescriptive (decision) analytics (Cont.) Decision trees Simulation-optimization Newsvendor problem and applications Decision rules, optimal statistical decisions Quality control, SPRT Evaluation analytics Learning analytics Decision psychology Heuristics and biases

Recommended readings Charniak (1991) (rest of paper) Build the network in Figure 2 www.aaai.org/ojs/index.php/aimagazine/article/viewFile/918/836 Pearl (2009) http://ftp.cs.ucla.edu/pub/stat_ser/r350.pdf Methods to Accelerate the Learning of Bayesian Network Structures, Daly and Shen (2007) https://pdfs.semanticscholar.org/e7d3/029e84a1775bb12e7e67541beaf2367f7a88.pdf Distinguishing cause from effect using observational data (Mooij et al., 2016), www.jmlr.org/papers/volume17/14-518/14-518.pdf Probabilistic computational causal discovery for systems biology (Lagani et al., 2016) www.mensxmachina.org/files/publications/Probabilistic%20Causal%20Discovery%20for%20Systems%20Biology_prePrint.pdf

Projects

Papers and projects: 3 types Applied: Analyze an application (description, prediction, causal analysis, decision, evaluation, learning) using high-value statistical consulting methods Research/develop software R packages, algorithms, CAT modules, etc. Research/review book or papers (3-5 articles) Explain a topic within statistical consulting Examples: Netica’s Bayesian inference algorithms, multicriteria decision-making, machine learning algorithms, etc.

Projects (cont.) Typical report paper is about 10-20 pages, font 12, space 1.5. (This is typical, not required) Content matters; length does not Typical in-class presentation is 20-30 minutes Can run longer if needed Purposes: Learn something interesting and useful; Either explain/show what you learned, or show how to use it in practice (or both)

Project proposals due March 17 If you have not yet done so, please send me a succinct description of what you want to do (and perhaps what you hope to learn by doing it). Problem to be addressed Methods to be researched/applied Hoped-for results Due by end of day on Friday, March 17th (though sooner is welcome) Key dates: April 14 for rough draft (or very good outline) Start in-class presentations/discussions April 18 May 4, 8:00 PM for final

Course schedule March 14: No class. (Work on project idea) March 17: Project/paper proposals due March 21: No class (Spring break) April 14: Draft of project/term paper due April 18, 25, May 2, (May 9): In-class presentations May 4: Final project/paper due by 8:00 PM

Prescriptive analytics (cont.)

Algorithms for optimizing actions Decision analysis framework: Choose act a from choice set A to maximize expected utility of consequence c, given a causal model c(a, s), Pr(s) or Pr(c | a, s), Pr(s) s = state = random variable = things that affect c other than the choice of act a Influence diagram algorithms Learning ID structure from data Validating causal mechanisms Using for inference and recommendations Simulation-optimization Robust optimization Adaptive optimization/learning algorithms

Prescriptive analytics methods Optimization Decision trees, Stochastic dynamic programming, optimal control Gittins indices Reinforcement learning (RL) algorithms Influence diagram solution algorithms Simulation-optimization Adaptive learning and optimization EVOP (Evolutionary operations) Multi-arm bandit problems, UCL strategies

Decision tree ingredients Three types of nodes Choice nodes (squares) Chance nodes (circles) Terminal nodes / value nodes Arcs show how decisions and chance events can unfold over time Uncertainties are resolved as time passes and choices are made

Solving decision trees “Backward induction” “Stochastic dynamic programming” “Average out and roll back”  implicitly, tree determines Pr(c | a) Procedure: Start at tips of tree, work backward Compute expected value at each chance node “Averaging out” Choose maximum expected value at each choice node

Obtaining Pr(s) from Decision trees http://www. eogogics                                                                      Decision 1: Develop or Do Not Develop Development Successful + Development Unsuccessful (70% X $172,000) + (30% x (- $500,000)) $120,400 + (-$150,000)

Obtaining Pr(s) from Decision trees http://www. eogogics                                                                      Decision 1: Develop or Do Not Develop Development Successful + Development Unsuccessful (70% X $172,000) + (30% x (- $500,000)) $120,400 + (-$150,000)

What happened to act a and state s. http://www. eogogics                                                                      Decision 1: Develop or Do Not Develop Development Successful + Development Unsuccessful (70% X $172,000) + (30% x (- $500,000)) $120,400 + (-$150,000)

What happened to act a and state s. http://www. eogogics                                                                      Decision 1: Develop or Do Not Develop Development Successful + Development Unsuccessful (70% X $172,000) + (30% x (- $500,000)) $120,400 + (-$150,000)

What happened to act a and state s. http://www. eogogics                                                                      What are the 3 possible acts in this tree?

What happened to act a and state s. http://www. eogogics                                                                      What are the 3 possible acts in this tree? (a) Don’t develop; (b) Develop, then rebuild if successful; (c) Develop, then new line if successful.

What happened to act a and state s. http://www. eogogics                                                                      Optimize decisions! What are the 3 possible acts in this tree? (a) Don’t develop; (b) Develop, then rebuild if successful; (c) Develop, then new line if successful.

Key points Solving decision trees (with decisions) requires embedded optimization Make future decisions optimally, given the information available when they are made Event trees = decision trees with no decisions Can be solved, to find outcome probabilities, by forward Monte-Carlo simulation, or by multiplication and addition In general, sequential decision-making cannot be modeled well using event trees. Must include (optimal choice | information)

What happened to state s. http://www. eogogics                                                                      What are the 4 possible states?

What happened to state s. http://www. eogogics                                                                      What are the 4 possible states? C1 can succeed or not; C2 can be high or low demand

Acts and states cause consequences http://www. eogogics                                                                     

Key theoretical insight A complex decision model can be viewed as a (possibly large) simple Pr(c | a) model. s = selection of branch at each chance node a = selection of branch at each choice node c = outcome at terminal node for (a, s) Pr(c | a) = sPr(c | a, s)*Pr(s) Other complex decision models can also be interpreted as c(a, s), Pr(c | a, s), or Pr(c |a) models s = system state & information signal a = decision rule (information  act) c may include changes in s and in possible a.

Real decision trees can quickly become “bushy messes” (Raiffa, 1968) with many duplicated sub-trees

Influence Diagrams help to avoid large trees http://en. wikipedia Often much more compact than decision trees

Limitations of decision trees Combinatorial explosion Example: Searching for a prize in one of N boxes or locations involves building a tree of depth N! = N(N – 1)…*2*1. Infinite trees Continuous variables When to stop growing a tree? How to evaluate utilities and probabilities?

Optimization formulations of decision problems Example: Prize is in location j with prior probability p(j), j = 1, 2, …, N It costs c(j) to inspect location j What search strategy minimizes expected cost of finding prize? What is a strategy? Order in which to inspect How many are there? N!

With two locations, 1 and 2 Strategy 1: Inspect 1, then 2 if needed: Expected cost: c1 + (1 – p1)c2 = c1 + c2 – p1c2 Strategy 2: Inspect 2, then 1 if needed: Expected cost: c2 + (1 – p2)c1 = c1 + c2 – p2c1 Strategy 1 has lower expected cost if: p1c2 > p2c1, or p1/c1 > p2/c2 So, look first at location with highest success probability per unit cost

With N locations Optimal decision rule: Always inspect next the (as-yet uninspected) location with the greatest success probability-to-cost ratio Example of an “index policy,” “Gittins index” If M players take turns, competing to find prize, each should still use this rule. A decision table or tree can be unwieldy even for such simple optimization problems

Other optimization formulations maxa A EU(a) Typically, a is a vector, A is the feasible set More generally, a is a strategy/policy/decision rule, A is the choice set of feasible strategies In previous example, A = set of permutations s.t. EU(a) = ∑cPr(c | a)u(c) Pr(c | a) = ∑sPr(c | a, s)p(s) g(a) ≤ 0 (feasible set, A)

Introduction to evaluation analytics

Evaluation analytics: How well are policies working? Algorithms for evaluating effects of actions, events, conditions Intervention analysis/interrupted time series Key idea: Compare predicted outcomes with no action to observed outcomes with it Counterfactual causal analysis Google’s new CausalImpact algorithm Quasi-experimental designs and analysis Refute non-causal explanations for data Compare to control groups to estimate effects

How did U.K. National Institute for Health and Clinical Excellence (NICE) recommendation of complete cessation of antibiotic prophylaxis for prevention of infective endocarditis in March, 2008 affect incidence of infective endocarditis? www.thelancet.com/journals/lancet/article/PIIS0140-6736(14)62007-9/fulltext?rss=yes

Different models yield different conclusions Different models yield different conclusions. So, how to deal with model uncertainty? Solution: Model ensembles, Bayesian Model Averaging (BMA) www.thelancet.com/journals/lancet/article/PIIS0140-6736(14)62007-9/fulltext?rss=yes

Nonlinear models complicate inference of intervention effects Solution: Non-parametric models, gradient boosting

Quasi-experiments: Refuting non-causal explanations with control groups Example: Do delinquency interventions work? http://www.slideshare.net/horatjitra/research-design-and-validity

Algorithms for evaluating effects of combinations of factors Classification trees Boosted trees, Random Forest, MARS Bayesian Network algorithms Discovery Conditional independence tests Validation Inference and explanation Response surface algorithms Adaptive learning, design of experiments

Learning analytics Learn to predict better Create ensemble of models, algorithms Use multiple machine learning algorithms Logistic regression, Random Forest, SVM, ANN, deep learning, gradient boosting, KNN, lasso, etc. “Stack” models (hybridize multiple predictions) Cross-validation assesses model performance Meta-learner combines performance-weighted predictors to produce an improved predictor Theoretical guarantees, practical successes (Kaggle competitions) Learn to decide better Low-regret learning of decision rules Theoretical guarantees (MDPs) practical performance http://www2.hawaii.edu/~chenx/ics699rl/grid/rl.html

Collaborative risk analytics: Multiple interacting learning agents http://groups.inf.ed.ac.uk/agents/index.php/Main/Projects

Collaborative risk analytics Global performance metrics Local information, control, tasks, priorities, rewards Hierarchical distributed control Collaborative sensing, filtering, deliberation, and decision-control networks of agents Mixed human and machine agents Autonomous agents vs. intelligent assistants http://www.cities.io/news/page/3/

Collaborative risk analytics: Games as labs for distributed AI Local information, control, tasks, priorities Hierarchical distributed control Collaborative sensing, deliberation, control networks From decentralized agents to effective risk analytics teams and HCI support Trust, reputation, performance Sharing information, attention, control, evaluation, learning http://people.idsia.ch/~juergen/learningrobots.html

Risk analytics toolkit: Summary Descriptive analytics Change-point analysis, likelihood ratio CPA Machine learning , response surfaces ML (LR, RF, GBM, SVM, ANN, KNN, etc.) Predictive analytics Bayesian networks, dynamic BN BN, DBN Bayesian model averaging BMA, ML Causal analytics & principles Causal BNs, systems dynamics (SD) DAGs, SD simulation Time series causation Prescriptive analytics: IDs, simulation-optimization, robust Evaluation analytics: QE, credit assignment, attribution Learning analytics Machine learning, superlearning ML Low-regret learning of decision rules Collaborative learning

Applied risk analytics toolkit: Toward more practical analytics Reorientation: From solving well-posed problems to discovering how to act more effectively Descriptive analytics: What’s happening? Predictive analytics: What’s coming next? Causal analytics: What can we do about it? Prescriptive analytics: What should we do? Evaluation analytics: How well is it working? Learning analytics: How to do better? Collaboration: How to do better together?