Uncertainty in AI.

Slides:



Advertisements
Similar presentations
Bayesian networks Chapter 14 Section 1 – 2. Outline Syntax Semantics Exact computation.
Advertisements

Probability: Review The state of the world is described using random variables Probabilities are defined over events –Sets of world states characterized.
Reasoning under Uncertainty: Marginalization, Conditional Prob., and Bayes Computer Science cpsc322, Lecture 25 (Textbook Chpt ) Nov, 6, 2013.
Bayesian Networks Chapter 14 Section 1, 2, 4. Bayesian networks A simple, graphical notation for conditional independence assertions and hence for compact.
1 Slides for the book: Probabilistic Robotics Authors: Sebastian Thrun Wolfram Burgard Dieter Fox Publisher: MIT Press, Web site for the book & more.
CPSC 322, Lecture 26Slide 1 Reasoning Under Uncertainty: Belief Networks Computer Science cpsc322, Lecture 27 (Textbook Chpt 6.3) March, 16, 2009.
Reasoning under Uncertainty: Conditional Prob., Bayes and Independence Computer Science cpsc322, Lecture 25 (Textbook Chpt ) March, 17, 2010.
… Hidden Markov Models Markov assumption: Transition model:
KI2 - 2 Kunstmatige Intelligentie / RuG Probabilities Revisited AIMA, Chapter 13.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Ai in game programming it university of copenhagen Welcome to... the Crash Course Probability Theory Marco Loog.
Uncertainty Chapter 13.
Review: Probability Random variables, events Axioms of probability
Bayesian Networks Textbook: Probabilistic Reasoning, Sections 1-2, pp
Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)
Soft Computing Lecture 17 Introduction to probabilistic reasoning. Bayesian nets. Markov models.
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Probabilistic Belief States and Bayesian Networks (Where we exploit the sparseness of direct interactions among components of a world) R&N: Chap. 14, Sect.
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
Probability and naïve Bayes Classifier Louis Oliphant cs540 section 2 Fall 2005.
Bayesian Statistics and Belief Networks. Overview Book: Ch 13,14 Refresher on Probability Bayesian classifiers Belief Networks / Bayesian Networks.
Introduction to Bayesian Networks
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
CSE PR 1 Reasoning - Rule-based and Probabilistic Representing relations with predicate logic Limitations of predicate logic Representing relations.
Reasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule Jim Little Uncertainty 2 Nov 3, 2014 Textbook §6.1.3.
Uncertainty. Assumptions Inherent in Deductive Logic-based Systems All the assertions we wish to make and use are universally true. Observations of the.
Reasoning under Uncertainty: Conditional Prob., Bayes and Independence Computer Science cpsc322, Lecture 25 (Textbook Chpt ) Nov, 5, 2012.
Marginalization & Conditioning Marginalization (summing out): for any sets of variables Y and Z: Conditioning(variant of marginalization):
Review: Probability Random variables, events Axioms of probability Atomic events Joint and marginal probability distributions Conditional probability distributions.
Uncertainty in AI. Birds can fly, right? Seems like common sense knowledge.
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Uncertainty Let action A t = leave for airport t minutes before flight Will A t get me there on time? Problems: 1.partial observability (road state, other.
1 Probability FOL fails for a domain due to: –Laziness: too much to list the complete set of rules, too hard to use the enormous rules that result –Theoretical.
Reasoning Under Uncertainty: Independence and Inference CPSC 322 – Uncertainty 5 Textbook §6.3.1 (and for HMMs) March 25, 2011.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
CSE 473 Uncertainty. © UW CSE AI Faculty 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one has stood.
Uncertainty & Probability CIS 391 – Introduction to Artificial Intelligence AIMA, Chapter 13 Many slides adapted from CMSC 421 (U. Maryland) by Bonnie.
Probabilistic Robotics Probability Theory Basics Error Propagation Slides from Autonomous Robots (Siegwart and Nourbaksh), Chapter 5 Probabilistic Robotics.
Matching ® ® ® Global Map Local Map … … … obstacle Where am I on the global map?                                   
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
Computer Science cpsc322, Lecture 25
Bayesian inference, Naïve Bayes model
Bayesian Networks Chapter 14 Section 1, 2, 4.
Quick Review Probability Theory
Quick Review Probability Theory
Today.
Read R&N Ch Next lecture: Read R&N
Probabilistic Reasoning
Reasoning Under Uncertainty: Conditioning, Bayes Rule & Chain Rule
Learning Bayesian Network Models from Data
Markov ó Kalman Filter Localization
Bayesian Networks Probability In AI.
Probabilistic Reasoning
Read R&N Ch Next lecture: Read R&N
Representing Uncertainty
CSE-490DF Robotics Capstone
Structure and Semantics of BN
CS 188: Artificial Intelligence Fall 2008
CAP 5636 – Advanced Artificial Intelligence
CS 188: Artificial Intelligence Fall 2007
Bayesian Statistics and Belief Networks
Probabilistic Reasoning
CS 188: Artificial Intelligence Fall 2007
CS 188: Artificial Intelligence
Structure and Semantics of BN
Read R&N Ch Next lecture: Read R&N
Read R&N Ch Next lecture: Read R&N
Chapter 14 February 26, 2004.
Presentation transcript:

Uncertainty in AI

Rule strengths or priorities in Expert Systems ...an old ad-hoc approach (with unclear semantics) penguin(x) →0.9 flies(x) bird(x) →0.5 flies(x) example: can assign 'salience' (integers) to defrules in Jess this will determine the order in which rules fire in the forward-chaining

Probability encode knowledge in the form of prior probabilities and conditional probabilities P(x speaks portugese|x is from Brazil)=0.9 P(x speaks portugese)=0.012 P(x is from Brazil)=0.007 P(x flies|x is a bird)=0.8 (?) inference is done by calculating posterior probabilities given evidence compute P(cavity | toothache, flossing, dental history, recent consumption of candy...) compute P(fed will raise interest rate | employment=5%, inflation=0.5%, GDP=2%, recent geopolitical events...)

Many modern knowledge-based systems are based on probabilistic inference including Bayesian networks, Hidden Markov Models, Markov Decision Problems example: Bayesian networks are used for inferring user goals or help needs from actions like mouse clicks in an automated software help system (think 'Clippy') Decision Theory combines utilities with probabilities of outcomes to decide actions to tak the challenge is capturing all the numbers needed for the prior and conditional probabilities objectivists (frequentists) - probabilities represent outcomes of trials/experiments subjectivists - probabilities are degrees of belief probability and statistics is at the core of many Machine Learning algorithms

Causal vs. diagnostic knowledge causal: P(x has a toothache|x has a cavity)=0.9 diagnostic: P(x has a cavity|x has a toothache)=0.5 typically it is easier to articulate knowledge in the causal direction, but we often want to use it in a diagnostic way to make inferences from observations

Bayes' Rule product rule : P(A,B) joint = P(A|B)*P(B) Bayes' Rule: convert between causal and diagnostic H = hypothesis (cause, disease) E = evidence (effect, symptoms)

Joint probability table (JPT) you can calculate answer to any question from JPT the problem is there are exponential # of entries (2N, where N is the number of binary random variables)

Joint probability table (JPT) you can calculate answer to any question from JPT the problem is there are exponential # of entries (2N, where N is the number of binary random variables) P(cavity | toothache)

Joint probability table (JPT) you can calculate answer to any question from JPT the problem is there are exponential # of entries (2N, where N is the number of binary random variables) P(cavity | toothache) = P(cavity  toothache) / P(toothache) = 0.016+0.064 / (0.108 + 0.012 + 0.016 + 0.064) = 0.4

Joint probability table (JPT) you can calculate answer to any question from JPT the problem is there are exponential # of entries (2N, where N is the number of binary random variables) P(cavity | toothache) = P(cavity  toothache) / P(toothache) = 0.016+0.064 / (0.108 + 0.012 + 0.016 + 0.064) = 0.4

P(A,B|C) = P(A|C)P(B|C) Solution to reduce complexity: Employ the Independence Assumption Most variables are not strictly independent, in that there is a statistical association (but which is cause? effect?). However, some many variables are conditionally independent. P(A,B|C) = P(A|C)P(B|C)

Bayesian networks note: absence of link between Burglary and JohnCalls means JohnCalls is conditionally independent of Burglary given Alarm

Equation for full joint probability

If some variables are unknown, you have to marginalize (sum) over them. There are many algorithms for trying to do efficient computation with Bayesian networks. For example: variable elimination factor graphs and message passing sampling

Lumiere (Microsoft) – Office Assistant

Hidden Markov Models transition probabilities and emission probabilities Viterbi algorithm for computing most probable state sequence from observations