CS 4/527: Artificial Intelligence

Slides:



Advertisements
Similar presentations
Utility theory U: O-> R (utility maps from outcomes to a real number) represents preferences over outcomes ~ means indifference We need a way to talk about.
Advertisements

Making Simple Decisions Chapter 16 Some material borrowed from Jean-Claude Latombe and Daphne Koller by way of Marie desJadines,
Making Simple Decisions
1 Decision Making and Utility Introduction –The expected value criterion may not be appropriate if the decision is a one-time opportunity with substantial.
Chapter 18 Statistical Decision Theory Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Statistics for Business and Economics 7 th.
Chapter 21 Statistical Decision Theory
DSC 3120 Generalized Modeling Techniques with Applications
Utility Axioms Axiom: something obvious, cannot be proven Utility axioms (rules for clear thinking)
Judgment and Decision Making in Information Systems Utility Functions, Utility Elicitation, and Risk Attitudes Yuval Shahar, M.D., Ph.D.
Cooperating Intelligent Systems Utility theory Chapter 16, AIMA.
91.420/543: Artificial Intelligence UMass Lowell CS – Fall 2010 Lecture 16: MEU / Utilities Oct 8, 2010 A subset of Lecture 8 slides of Dan Klein – UC.
Making Simple Decisions Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 16.
CS 188: Artificial Intelligence Fall 2009 Lecture 17: Bayes Nets IV 10/27/2009 Dan Klein – UC Berkeley.
Decision Making Under Uncertainty Russell and Norvig: ch 16 CMSC421 – Fall 2006.
CS 188: Artificial Intelligence Fall 2009 Lecture 8: MEU / Utilities 9/22/2009 Dan Klein – UC Berkeley Many slides over the course adapted from either.
CMSC 671 Fall 2003 Class #26 – Wednesday, November 26 Russell & Norvig 16.1 – 16.5 Some material borrowed from Jean-Claude Latombe and Daphne Koller by.
PGM 2003/04 Tirgul7 Foundations of Decision Theory (mostly from Pearl)
Decision Analysis (cont)
Expectimax Evaluation
Reminder Midterm Mar 7 Project 2 deadline Mar 18 midnight in-class
Quiz 5: Expectimax/Prob review
Making Simple Decisions
Making Simple Decisions Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 16.
CS 188: Artificial Intelligence Spring 2007 Lecture 21:Reinforcement Learning: I Utilities and Simple decisions 4/10/2007 Srini Narayanan – ICSI and UC.
Decision theory under uncertainty
Decision Making Under Uncertainty CMSC 671 – Fall 2010 R&N, Chapters , , material from Lise Getoor, Jean-Claude Latombe, and.
Making Simple Decisions Utility Theory MultiAttribute Utility Functions Decision Networks The Value of Information Summary.
Models for Strategic Marketing Decision Making. Market Entry Decisions To enter first or to wait Sources of First-Mover Advantages –Technological leadership.
Chapter 16: Making Simple Decision March 23, 2004.
CS 188: Artificial Intelligence Fall 2008 Lecture 7: Expectimax Search 9/18/2008 Dan Klein – UC Berkeley Many slides over the course adapted from either.
CS 188: Artificial Intelligence Fall 2007 Lecture 8: Expectimax Search 9/20/2007 Dan Klein – UC Berkeley Many slides over the course adapted from either.
Web-Mining Agents Agents and Rational Behavior Decision-Making under Uncertainty Simple Decisions Ralf Möller Universität zu Lübeck Institut für Informationssysteme.
Making Simple Decisions Chapter 16 Some material borrowed from Jean-Claude Latombe and Daphne Koller by way of Marie desJadines,
CS 188: Artificial Intelligence Spring 2009 Lecture 20: Decision Networks 4/2/2009 John DeNero – UC Berkeley Slides adapted from Dan Klein.
Chapter 16 March 25, Probability Theory: What an agent should believe based on the evidence Utility Theory: What the agent wants Decision Theory:
CS 188: Artificial Intelligence Fall 2006 Lecture 8: Expectimax Search 9/21/2006 Dan Klein – UC Berkeley Many slides over the course adapted from either.
Decision Making ECE457 Applied Artificial Intelligence Spring 2007 Lecture #10.
Chapter 19 Statistical Decision Theory ©. Framework for a Decision Problem action i.Decision maker has available K possible courses of action : a 1, a.
CS 188: Artificial Intelligence Fall 2008 Lecture 8: MEU / Utilities 9/23/2008 Dan Klein – UC Berkeley Many slides over the course adapted from either.
CS 188: Artificial Intelligence Spring 2006
Vincent Conitzer CPS Utility theory Vincent Conitzer
Utilities and Decision Theory
Nevin L. Zhang Room 3504, phone: ,
Ralf Möller Universität zu Lübeck Institut für Informationssysteme
Making Simple Decisions
CHAPTER 1 FOUNDATIONS OF FINANCE I: EXPECTED UTILITY THEORY
ECE457 Applied Artificial Intelligence Fall 2007 Lecture #10
ECE457 Applied Artificial Intelligence Spring 2008 Lecture #10
Expectimax Lirong Xia. Expectimax Lirong Xia Project 2 MAX player: Pacman Question 1-3: Multiple MIN players: ghosts Extend classical minimax search.
CS 4/527: Artificial Intelligence
Risk Chapter 11.
CS 188: Artificial Intelligence Fall 2009
Game playing.
Statistical Decision Theory
Rational Decisions and
L03 Utility.
L03 Utility.
Utility Theory Decision Theory.
13. Acting under Uncertainty Wolfram Burgard and Bernhard Nebel
Choices, Values and Frames
Chapter 17 Decision Making
Expectimax Lirong Xia.
Making Simple Decisions
Behavioral Finance Economics 437.
Vincent Conitzer Utility theory Vincent Conitzer
CS 188: Artificial Intelligence Fall 2007
Utilities and Decision Theory
Utilities and Decision Theory
CS51A David Kauchak Spring 2019
Making Simple Decisions
Presentation transcript:

CS 4/527: Artificial Intelligence Bayes Nets - Sampling and Decision Networks Please retain proper attribution and the reference to ai.berkeley.edu. Thanks! Instructor: Jared Saia University of New Mexico

Utilities

Maximum Expected Utility Principle of maximum expected utility: A rational agent should chose the action that maximizes its expected utility, given its knowledge Questions: Where do utilities come from? How do we know such utilities even exist? How do we know that averaging even makes sense? What if our behavior (preferences) can’t be described by utilities?

The need for numbers x2 40 20 30 1600 400 900 For worst-case minimax reasoning, terminal value scale doesn’t matter We just want better states to have higher evaluations (get the ordering right) The optimal decision is invariant under any monotonic transformation For average-case expectimax reasoning, we need magnitudes to be meaningful

Utilities Utilities are functions from outcomes (states of the world) to real numbers that describe an agent’s preferences Where do utilities come from? In a game, may be simple (+1/-1) Utilities summarize the agent’s goals Theorem: any “rational” preferences can be summarized as a utility function We hard-wire utilities and let behaviors emerge Why don’t we let agents pick utilities? Why don’t we prescribe behaviors?

Utilities: Uncertain Outcomes Getting ice cream Get Single Get Double Oops Whew!

Preferences A Prize A Lottery A A B An agent must have preferences among: Prizes: A, B, etc. Lotteries: situations with uncertain prizes L = [p, A; (1-p), B] Notation: Preference: A > B Indifference: A ~ B A p 1-p A B

Rationality

Rational Preferences We want some constraints on preferences before we call them rational, such as: For example: an agent with intransitive preferences can be induced to give away all of its money If B > C, then an agent with C would pay (say) 1 cent to get B If A > B, then an agent with B would pay (say) 1 cent to get A If C > A, then an agent with A would pay (say) 1 cent to get C Axiom of Transitivity: (A > B)  (B > C)  (A > C) Apple, banana, carrot

The Axioms of Rationality Rational Preferences The Axioms of Rationality Orderability: (A > B)  (B > A)  (A ~ B) Transitivity: (A > B)  (B > C)  (A > C) Continuity: (A > B > C)  p [p, A; 1-p, C] ~ B Substitutability: (A ~ B)  [p, A; 1-p, C] ~ [p, B; 1-p, C] Monotonicity: (A > B)  (p  q)  [p, A; 1-p, B]  [q, A; 1-q, B] Theorem: Rational preferences imply behavior describable as maximization of expected utility

MEU Principle U([p1,S1; … ; pn,Sn]) = p1U(S1) + … + pnU(Sn) Theorem [Ramsey, 1931; von Neumann & Morgenstern, 1944] Given any preferences satisfying these constraints, there exists a real-valued function U such that: U(A)  U(B)  A  B U([p1,S1; … ; pn,Sn]) = p1U(S1) + … + pnU(Sn) I.e. values assigned by U preserve preferences of both prizes and lotteries! Optimal policy invariant under positive affine transformation U’ = aU+b, a>0 Maximum expected utility (MEU) principle: Choose the action that maximizes expected utility Note: rationality does not require representing or manipulating utilities and probabilities E.g., a lookup table for perfect tic-tac-toe

Human Utilities

Human Utilities Pay $50 No change Instant death Utilities map states to real numbers. Which numbers? Standard approach to assessment (elicitation) of human utilities: Compare a prize A to a standard lottery Lp between “best possible prize” uT with probability p “worst possible catastrophe” u with probability 1-p Adjust lottery probability p until indifference: A ~ Lp Resulting p is a utility in [0,1] Pay $50 0.999999 0.000001 No change Instant death

Money Money does not behave as a utility function, but we can talk about the utility of having money (or being in debt) Given a lottery L = [p, $X; (1-p), $Y] The expected monetary value EMV(L) = pX + (1-p)Y The utility is U(L) = pU($X) + (1-p)U($Y) Typically, U(L) < U( EMV(L) ) In this sense, people are risk-averse E.g., how much would you pay for a lottery ticket L=[0.5, $10,000; 0.5, $0]? The certainty equivalent of a lottery CE(L) is the cash amount such that CE(L) ~ L The insurance premium is EMV(L) - CE(L) If people were risk-neutral, this would be zero! Draw Jensen’s for utility curve

Post-decision Disappointment: the Optimizer’s Curse Usually we don’t have direct access to exact utilities, only estimates E.g., you could make one of k investments An unbiased expert assesses their expected net profit V1,…,Vk You choose the best one V* With high probability, its actual value is considerably less than V* This is a serious problem in many areas: Future performance of mutual funds Efficacy of drugs measured by trials Statistical significance in scientific papers Winning an auction Suppose true net profit is 0 and estimate ~ N(0,1); Max of k estimates:

Decision Networks

Decision Networks Umbrella U Weather Forecast There exists a ghostbusters demo Forecast

Decision Networks Decision network = Bayes net + Actions + Utilities Action nodes (rectangles, cannot have parents, will have value fixed by algorithm) Utility nodes (diamond, depends on action and chance nodes) Decision algorithm: Fix evidence e For each possible action a Fix action node to a Compute posterior P(W|e,a) for parents W of U Compute expected utility w P(w|e,a) U(a,w) Return action with highest expected utility Umbrella U Weather Forecast Bayes net inference! There exists a ghostbusters demo

Example: Take an umbrella? Decision algorithm: Fix evidence e For each possible action a Fix action node to a Compute posterior P(W|e,a) for parents W of U Compute expected utility of action a: w P(w|e,a) U(a,w) Return action with highest expected utility A W U(A,W) leave sun 100 rain take 20 70 Bayes net inference! Umbrella U W P(W) sun 0.7 Umbrella = leave EU(leave|F=bad) = w P(w|F=bad) U(leave,w) Weather = 0.34x100 + 0.66x0 = 34 W P(W|F=bad) sun 0.34 rain 0.66 Umbrella = take EU(take|F=bad) = w P(w|F=bad) U(take,w) = 0.34x20 + 0.66x70 = 53 Forecast =bad W P(F=bad|W) sun 0.17 rain 0.77 Optimal decision = take!