1 Probability FOL fails for a domain due to: –Laziness: too much to list the complete set of rules, too hard to use the enormous rules that result –Theoretical.

Slides:



Advertisements
Similar presentations
Bayesian networks Chapter 14 Section 1 – 2. Outline Syntax Semantics Exact computation.
Advertisements

Artificial Intelligence Universitatea Politehnica Bucuresti Adina Magda Florea
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
Reasoning under Uncertainty Department of Computer Science & Engineering Indian Institute of Technology Kharagpur.
1 Knowledge Engineering for Bayesian Networks. 2 Probability theory for representing uncertainty l Assigns a numerical degree of belief between 0 and.
Identifying Conditional Independencies in Bayes Nets Lecture 4.
Bayesian Networks CPSC 386 Artificial Intelligence Ellen Walker Hiram College.
1 22c:145 Artificial Intelligence Bayesian Networks Reading: Ch 14. Russell & Norvig.
Bayesian Networks Chapter 14 Section 1, 2, 4. Bayesian networks A simple, graphical notation for conditional independence assertions and hence for compact.
Uncertain knowledge and reasoning
Review: Bayesian learning and inference
Bayesian Networks. Motivation The conditional independence assumption made by naïve Bayes classifiers may seem to rigid, especially for classification.
Probabilistic Reasoning Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 14 (14.1, 14.2, 14.3, 14.4) Capturing uncertain knowledge Probabilistic.
Bayesian networks Chapter 14 Section 1 – 2.
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 14 Jim Martin.
Bayesian Belief Networks
1 Bayesian Reasoning Chapter 13 CMSC 471 Adapted from slides by Tim Finin and Marie desJardins.
Intro to AI Uncertainty Ruth Bergman Fall Why Not Use Logic? Suppose I want to write down rules about medical diagnosis: Diagnostic rules: A x has(x,sorethroat)
University College Cork (Ireland) Department of Civil and Environmental Engineering Course: Engineering Artificial Intelligence Dr. Radu Marinescu Lecture.
Bayesian Networks Russell and Norvig: Chapter 14 CMCS424 Fall 2003 based on material from Jean-Claude Latombe, Daphne Koller and Nir Friedman.
Belief Networks Russell and Norvig: Chapter 15 CS121 – Winter 2002.
Bayesian Reasoning. Tax Data – Naive Bayes Classify: (_, No, Married, 95K, ?)
1 Probabilistic Belief States and Bayesian Networks (Where we exploit the sparseness of direct interactions among components of a world) R&N: Chap. 14,
Bayesian Networks Russell and Norvig: Chapter 14 CMCS421 Fall 2006.
Bayesian networks More commonly called graphical models A way to depict conditional independence relationships between random variables A compact specification.
Uncertainty Chapter 13.
EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS
Bayes Nets and Probabilities
CSCI 121 Special Topics: Bayesian Network Lecture #1: Reasoning Under Uncertainty.
Read R&N Ch Next lecture: Read R&N
Bayesian networks Chapter 14. Outline Syntax Semantics.
Uncertain KR&R Chapter 9. 2 Outline Probability Bayesian networks Fuzzy logic.
Bayesian networks Chapter 14 Section 1 – 2. Bayesian networks A simple, graphical notation for conditional independence assertions and hence for compact.
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Probabilistic Belief States and Bayesian Networks (Where we exploit the sparseness of direct interactions among components of a world) R&N: Chap. 14, Sect.
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
Bayesian Statistics and Belief Networks. Overview Book: Ch 13,14 Refresher on Probability Bayesian classifiers Belief Networks / Bayesian Networks.
1 Monte Carlo Artificial Intelligence: Bayesian Networks.
Introduction to Bayesian Networks
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
CSE PR 1 Reasoning - Rule-based and Probabilistic Representing relations with predicate logic Limitations of predicate logic Representing relations.
Chapter 13 February 19, Acting Under Uncertainty Rational Decision – Depends on the relative importance of the goals and the likelihood of.
Uncertainty. Assumptions Inherent in Deductive Logic-based Systems All the assertions we wish to make and use are universally true. Observations of the.
Probabilistic Reasoning [Ch. 14] Bayes Networks – Part 1 ◦Syntax ◦Semantics ◦Parameterized distributions Inference – Part2 ◦Exact inference by enumeration.
Uncertainty Chapter 13. Outline Uncertainty Probability Syntax and Semantics Inference Independence and Bayes' Rule.
Review: Bayesian inference  A general scenario:  Query variables: X  Evidence (observed) variables and their values: E = e  Unobserved variables: Y.
Uncertainty Let action A t = leave for airport t minutes before flight Will A t get me there on time? Problems: 1.partial observability (road state, other.
Conditional Probability, Bayes’ Theorem, and Belief Networks CISC 2315 Discrete Structures Spring2010 Professor William G. Tanner, Jr.
CSE 473 Uncertainty. © UW CSE AI Faculty 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one has stood.
Artificial Intelligence Universitatea Politehnica Bucuresti Adina Magda Florea
Belief Networks CS121 – Winter Other Names Bayesian networks Probabilistic networks Causal networks.
PROBABILISTIC REASONING Heng Ji 04/05, 04/08, 2016.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
Intro to AI Uncertainty Ruth Bergman Fall Why Not Use Logic? Suppose I want to write down rules about medical diagnosis: Diagnostic rules:  p Symptom(p,Toothache)
CS 2750: Machine Learning Directed Graphical Models
Bayesian Networks Chapter 14 Section 1, 2, 4.
Bayesian networks Chapter 14 Section 1 – 2.
Presented By S.Yamuna AP/CSE
Uncertain KR&R Chapter 9.
Uncertainty Chapter 13.
Conditional Probability, Bayes’ Theorem, and Belief Networks
Uncertainty in AI.
Bayesian Statistics and Belief Networks
Belief Networks CS121 – Winter 2003 Belief Networks.
Bayesian networks Chapter 14 Section 1 – 2.
Probabilistic Reasoning
Bayesian Networks CSE 573.
Representing Uncertainty
Presentation transcript:

1 Probability FOL fails for a domain due to: –Laziness: too much to list the complete set of rules, too hard to use the enormous rules that result –Theoretical ignorance: there is no complete theory for the domain –Practical ignorance: have not or cannot run all necessary tests

2 Probability Probability = a degree of belief Probability comes from: –Frequentist: experiments and statistical assessment –Objectivist: real aspects of the universe –Subjectivist: a way of characterizing an agent’s beliefs Decision theory = probability + utility theory

3 Probability Prior probability: probability in the absence of any other information P(Dice = 2) = 1/6 random variable: Dice domain = probability distribution: P(Dice) =

4 Probability Conditional probability: probability in the presence of some evidence P(Dice = 2 | Dice is even) = 1/3 P(Dice = 2 | Dice is odd) = 0 P(A | B) = P(A  B)/P(B) P(A  B) = P(A | B).P(B)

5 Probability Example: S = stiff neck M = meningitis P(S | M) = 0.5 P(M) = 1/50000 P(S) = 1/20  P(M | S) = P(S | M).P(M)/P(S) = 1/5000

6 Probability Joint probability distributions: X: Y: P(X = x i, Y = y j )

7 Probability Axioms: 0  P(A)  1 P(true) = 1 and P(false) = 0 P(A  B) = P(A) + P(B)  P(A  B)

8 Probability Derived properties: P(  A) = 1  P(A) P(U) = P(A 1 ) + P(A 2 ) P(A n ) U = A 1  A 2 ...  A n collectively exhaustive A i  A j = falsemutually exclusive

9 Probability Bayes’ theorem: P(H i | E) = P(H i  E)/P(E) = P(E | H i ).P(H i )/  P(E | H i ).P(H i ) H i ’s are collectively exhaustive & mutually exclusive

10 Probability Problem: a full joint probability distribution P(X 1, X 2,..., X n ) is sufficient for computing any probability on X i 's, but the number of joint probabilities is exponential.

11 Probability Independence: P(A  B) = P(A).P(B) P(A) = P(A | B) Conditional independence: P(A  B | E) = P(A | E).P(B | E) P(A | E) = P(A | E  B)

12 Probability Example: P(Toothache | Cavity  Catch) = P(Toothache | Cavity) P(Catch | Cavity  Toothache) = P(Catch | Cavity)

13 Probability "In John's and Mary's house, an alarm is installed to sound in case of burglary or earthquake. When the alarm sounds, John and Mary may make a call for help or rescue."

14 Probability "In John's and Mary's house, an alarm is installed to sound in case of burglary or earthquake. When the alarm sounds, John and Mary may make a call for help or rescue." Q1: If earthquake happens, how likely will John make a call?

15 Probability "In John's and Mary's house, an alarm is installed to sound in case of burglary or earthquake. When the alarm sounds, John and Mary may make a call for help or rescue." Q1: If earthquake happens, how likely will John make a call? Q2: If the alarm sounds, how likely is the house burglarized?

16 Bayesian Networks Pearl, J. (1982). Reverend Bayes on Inference Engines: A Distributed Hierarchical Approach, presented at the Second National Conference on Artificial Intelligence (AAAI-82), Pittsburgh, Pennsylvania 2000 AAAI Classic Paper Award

17 Bayesian Networks BurglaryEarthquake Alarm JohnCalls MaryCalls P(B).001 P(E).002 B EP(A) T T F F T F AP(J) TFTF AP(M) TFTF.70.01

18 Bayesian Networks Syntax: –A set of random variables making up the nodes –A set of directed links connecting pairs of nodes –Each node has a conditional probability table that quantifies the effects of its parent nodes –The graph has no directed cycles A B C D A B C D yes no

19 Bayesian Networks Semantics: Ordering the nodes: X i is a predecessor of X j  i < j P(X 1, X 2, …, X n ) = P(X n | X n-1, …, X 1 ).P(X n-1 | X n-2, …, X 1 ). ….P(X 2 | X 1 ).P(X 1 ) =  i P(X i | X i-1, …, X 1 ) =  i P(X i | Parents(X i )) P (X i | X i-1, …, X 1 ) = P(X i | Parents(X i )) Parents(X i )  {X i-1, …, X 1 } Each node is conditionally independent of its predecessors given its parents

20 Bayesian Networks Example: P(J  M  A   B   E) = P(J | A).P(M | A).P(A |  B   E).P(  B).P(  E)

21 Bayesian Networks P(Query | Evidence) = ? –Diagnostic (from effects to causes): P(B | J) –Causal (from causes to effects): P(J | B) –Intercausal (between causes of a common effect): P(B | A, E) –Mixed: P(A | J,  E), P(B | J,  E)

22 P(B | A) = P(B  A)/P(A) =  P(B  A) Bayesian Networks

23 Bayesian Networks Why Bayesian Networks?

24 General Conditional Independence Y1Y1 YnYn U1U1 UmUm Z 1j Z nj X A node (X) is conditionally independent of its non-descendents (Z ij 's), given its parents (U i 's)

25 General Conditional Independence Y1Y1 YnYn U1U1 UmUm Z 1j Z nj X A node (X) is conditionally independent of all other nodes, given its parents (U i 's), children (Y i 's), and children's parents (Z ij 's)

26 General Conditional Independence X and Y are conditionally independent given E Z Z Z XYE

27 General Conditional Independence Example: Battery GasIgnitionRadio Starts Moves

28 General Conditional Independence Example: Battery GasIgnitionRadio Starts Moves Gas - (no evidence at all) - Radio Gas  Radio | Starts Gas  Radio | Moves

29 Network Construction Wrong ordering: –Produces a more complicated net –Requires more probabilities to be specified –Produces slight relationships that require difficult and unnatural probability judgments –Fails to represent all the conditional independence relationships  a lot of unnecessary probabilities

30 Network Construction Models: –Diagnostic: symptoms to causes –Causal: causes to symptoms Add "root causes" first, then the variables they influence  fewer and more comprehensible probabilities

31 Dempster-Shafer Theory Example: disease diagnostics D = {Allergy, Flu, Cold, Pneumonia}

32 Dempster-Shafer Theory Example: disease diagnostics D = {Allergy, Flu, Cold, Pneumonia} Basic probability assignment: m: 2 D  [0, 1] m 1 : {Flu, Cold, Pneu} 0.6 D 0.4

33 Dempster-Shafer Theory Belief: Bel: 2 D  [0, 1] Bel(S) =  X  S m(X)

34 Dempster-Shafer Theory Belief: Bel: 2 D  [0, 1] Bel(S) =  X  S m(X) Plausibility: Pl(p) = 1  Bel(  p) p: [Bel(p), Pl(p)]

35 Dempster-Shafer Theory Ignorance: Bel(p) + Bel(  p)  1

36 Dempster-Shafer Theory Combination rules: m 1 and m 2  X  Y=Z m 1 (X).m 2 (Y) m(Z) = 1   X  Y=  m 1 (X).m 2 (Y)

37 Exercises In Russell’s AIMA: Chapters