Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Lecture 30 of 42 William H. Hsu Department.

Slides:



Advertisements
Similar presentations
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Advertisements

IMPORTANCE SAMPLING ALGORITHM FOR BAYESIAN NETWORKS
Artificial Intelligence Chapter 19 Reasoning with Uncertain Information Biointelligence Lab School of Computer Sci. & Eng. Seoul National University.
Bayes Nets Rong Jin. Hidden Markov Model  Inferring from observations (o i ) to hidden variables (q i )  This is a general framework for representing.
Computing & Information Sciences Kansas State University Lecture 11 of 42 CIS 530 / 730 Artificial Intelligence Lecture 11 of 42 William H. Hsu Department.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Introduction to Graphical Models.
Part 2 of 3: Bayesian Network and Dynamic Bayesian Network.
1 Bayesian Reasoning Chapter 13 CMSC 471 Adapted from slides by Tim Finin and Marie desJardins.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Monday, March 6, 2000 William.
Kansas State University Department of Computing and Information Sciences Kansas State University KDD Lab ( Graphical.
Computing & Information Sciences Kansas State University Lecture 28 of 42 CIS 530 / 730 Artificial Intelligence Lecture 28 of 42 William H. Hsu Department.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 26 of 41 Friday, 22 October.
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
Computing & Information Sciences Kansas State University Lecture 27 of 42 CIS 530 / 730 Artificial Intelligence Lecture 27 of 42 William H. Hsu Department.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Naive Bayes Classifier
Kansas State University Department of Computing and Information Sciences Kansas State University KDD Lab ( Permutation.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 25 Wednesday, 20 October.
Computing & Information Sciences Kansas State University Wednesday, 15 Oct 2008CIS 530 / 730: Artificial Intelligence Lecture 20 of 42 Wednesday, 15 October.
Computing & Information Sciences Kansas State University Monday, 29 Oct 2008CIS 530 / 730: Artificial Intelligence Lecture 25 of 42 Wednesday, 29 October.
Introduction to Bayesian Networks
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
Computing & Information Sciences Kansas State University Wednesday, 22 Oct 2008CIS 530 / 730: Artificial Intelligence Lecture 22 of 42 Wednesday, 22 October.
Computing & Information Sciences Kansas State University Wednesday, 20 Sep 2006CIS 490 / 730: Artificial Intelligence Lecture 12 of 42 Wednesday, 20 September.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 28 of 41 Friday, 22 October.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Friday, 29 October 2004 William.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Friday, March 10, 2000 William.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 25 of 41 Monday, 25 October.
Reasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule Jim Little Uncertainty 2 Nov 3, 2014 Textbook §6.1.3.
Computing & Information Sciences Kansas State University Wednesday, 25 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 26 of 42 Wednesday. 25 October.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 11 of 41 Wednesday, 15.
Computing & Information Sciences Kansas State University Lecture 8 of 42 CIS 530 / 730 Artificial Intelligence Lecture 8 of 42 William H. Hsu Department.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 13 of 41 Monday, 20 September.
Computing & Information Sciences Kansas State University Lecture 13 of 42 CIS 530 / 730 Artificial Intelligence Lecture 13 of 42 William H. Hsu Department.
Computing & Information Sciences Kansas State University Lecture 40 of 42 CIS 530 / 730 Artificial Intelligence Lecture 40 of 42 A Brief Survey of Computer.
Computing & Information Sciences Kansas State University Data Sciences Summer Institute Multimodal Information Access and Synthesis Learning and Reasoning.
Computing & Information Sciences Kansas State University Monday, 06 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 31 of 42 Monday, 06 November.
Chapter 6 Bayesian Learning
Computing & Information Sciences Kansas State University Lecture 14 of 42 CIS 530 / 730 Artificial Intelligence Lecture 14 of 42 William H. Hsu Department.
Computing & Information Sciences Kansas State University Monday, 25 Sep 2006CIS 490 / 730: Artificial Intelligence Lecture 14 of 42 Monday, 25 September.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 14 of 41 Wednesday, 22.
CPSC 422, Lecture 11Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Oct, 2, 2015.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Wednesday, 21 February 2007.
Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 27 of 42 Friday, 27 October.
Computing & Information Sciences Kansas State University Monday, 23 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 25 of 42 Monday, 23 October.
Computing & Information Sciences Kansas State University Friday, 20 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 24 of 42 Friday, 20 October.
Computing & Information Sciences Kansas State University Lecture 12 of 42 CIS 530 / 730 Artificial Intelligence Lecture 12 of 42 William H. Hsu Department.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Graphical Models of Probability.
Computing & Information Sciences Kansas State University Wednesday, 08 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 32 of 42 Wednesday, 08 November.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 24 of 41 Monday, 18 October.
Bayesian Learning Bayes Theorem MAP, ML hypotheses MAP learners
Computing & Information Sciences Kansas State University Wednesday, 08 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 32 of 42 Wednesday, 08 November.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Wednesday, 28 February 2007.
Computing & Information Sciences Kansas State University Monday, 09 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 19 of 42 Monday, 09 October.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 14 of 42 Wednesday, 22.
Computing & Information Sciences Kansas State University Wednesday, 04 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 17 of 42 Wednesday, 04 October.
Computing & Information Sciences Kansas State University Friday, 03 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 30 of 42 Friday, 03 November.
Computing & Information Sciences Kansas State University Friday, 13 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 21 of 42 Friday, 13 October.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Monday, 01 February 2016 William.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
Naive Bayes Classifier. REVIEW: Bayesian Methods Our focus this lecture: – Learning and classification methods based on probability theory. Bayes theorem.
Computing & Information Sciences Kansas State University Wednesday, 25 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 26 of 42 Wednesday. 25 October.
Computing & Information Sciences Kansas State University Monday, 18 Sep 2006CIS 490 / 730: Artificial Intelligence Lecture 11 of 42 Monday, 18 September.
Bayesian Decision Theory Introduction to Machine Learning (Chap 3), E. Alpaydin.
Computing & Information Sciences Kansas State University Friday, 31 Oct 2008CIS 530 / 730: Artificial Intelligence Lecture 26 of 42 Friday, 31 October.
Computing & Information Sciences Kansas State University Wednesday, 01 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 29 of 42 Wednesday, 01 November.
Reasoning Under Uncertainty: Conditioning, Bayes Rule & Chain Rule
Class #21 – Monday, November 10
Naive Bayes Classifier
Presentation transcript:

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Lecture 30 of 42 William H. Hsu Department of Computing and Information Sciences, KSU KSOL course page: Course web site: Instructor home page: Reading for Next Class: Kevin Murphy‘s survey on BNs, learning: BNJ homepage: Reasoning under Uncertainty: Inference and Software Tools, Part 2 of 2 Discussion: BNJ

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Lecture Outline Reading for Next Class: Sections 14.1 – 14.2 (p. 492 – 499), R&N 2 e Last Class: Uncertainty, Probability, 13 (p ), R&N 2 e  Where uncertainty is encountered: reasoning, planning, learning (later)  Sources: sensor error, incomplete/inaccurate domain theory, randomness Today: Probability Intro, Continued, Chapter 13, R&N 2 e  Why probability  Axiomatic basis: Kolmogorov  With utility theory: sound foundation of rational decision making  Joint probability  Independence  Probabilistic reasoning: inference by enumeration  Conditioning  Bayes’s theorem (aka Bayes’ rule)  Conditional independence Coming Week: More Applied Probability, Graphical Models

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Acknowledgements © Russell, S. J. University of California, Berkeley Norvig, P. Slides from: Peter Norvig Director of Research Google Stuart J. Russell Professor of Computer Science Chair, Department of Electrical Engineering and Computer Sciences Smith-Zadeh Prof. in Engineering University of California - Berkeley Lotfali Asker-Zadeh (Lotfi A. Zadeh) Professor of Computer Science Department of Electrical Engineering and Computer Sciences Director, Berkeley Initiative in Soft Computing University of California - Berkeley © 2008 Zadeh, L. A. University of California, Berkeley

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Sample Space (  ): Range of Random Variable X Probability Measure Pr(  )   denotes range of observations; X:   Probability Pr, or P: measure over power set 2  - event space  In general sense, Pr(X = x   ) is measure of belief in X = x  P(X = x) = 0 or P(X = x) = 1: plain (aka categorical) beliefs  Can’t be revised; all other beliefs are subject to revision Kolmogorov Axioms  1.  x  . 0  P(X = x)  1  2. P(  )   x   P(X = x) = 1  3. Joint Probability: P(X 1  X 2 )  Prob. of Joint Event X 1  X 2 Independence: P(X 1  X 2 ) = P(X 1 )  P(X 2 ) Probability: Basic Definitions and Axioms

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence fuzzy set Non-Probabilistic Representation [1]: Concept of Fuzzy Set Adapted from slide © 2008 L. A. Zadeh, UC Berkeleyhttp://bit.ly/39shSQhttp://bit.ly/39shSQ class set precisiation unprecisiated generalization Informally, a fuzzy set, A, in a universe of discourse, U, is a class with a fuzzy boundary.

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Non-Probabilistic Representation [2]: Precisiation & Degree of Membership Adapted from slide © 2008 L. A. Zadeh, UC Berkeleyhttp://bit.ly/39shSQhttp://bit.ly/39shSQ Set A in U: Class with Crisp Boundary Precisiation: Association with Function whose Domain is U Precisiation of Crisp Sets  Through association with (Boolean-valued) characteristic function  c A : U  {0, 1} Precisiation of Fuzzy Sets  Through association with membership function  µ A : U  [0, 1]  µ A (u), u  U, represents grade of membership of u in A Degree of Membership  Membership in A: matter of degree  “In fuzzy logic everything is or is allowed to be a matter of degree.” – Zadeh

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Non-Probabilistic Representation [3]: Fuzzy Set Example – Middle-Age Adapted from slide © 2008 L. A. Zadeh, UC Berkeleyhttp://bit.ly/39shSQhttp://bit.ly/39shSQ μ 1 0 definitely middle-age definitely not middle-age middle-age core of middle-age “Linguistic” Variables: Qualitative, Based on Descriptive Terms Imprecision of Meaning = Elasticity of Meaning Elasticity of Meaning = Fuzziness of Meaning

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Product Rule (Alternative Statement of Bayes’s Theorem)  Proof: requires axiomatic set theory, as does Bayes’s Theorem Sum Rule  Sketch of proof (immediate from axiomatic set theory)  Draw a Venn diagram of two sets denoting events A and B  Let A  B denote the event corresponding to A  B… Theorem of Total Probability  Suppose events A 1, A 2, …, A n are mutually exclusive and exhaustive  Mutually exclusive: i  j  A i  A j =   Exhaustive:  P(A i ) = 1  Then  Proof: follows from product rule and 3 rd Kolmogorov axiom A B Basic Formulas For Probabilities

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Theorem P(h)  Prior Probability of Assertion (Hypothesis) h  Measures initial beliefs (BK) before any information is obtained (hence prior) P(D)  Prior Probability of Data (Observations) D  Measures probability of obtaining sample D (i.e., expresses D) P(h | D)  Probability of h Given D  | denotes conditioning - hence P(h | D) conditional (aka posterior) probability P(D | h)  Probability of D Given h  Measures probability of observing D when h is correct (“generative” model) P(h  D)  Joint Probability of h and D  Measures probability of observing D and of h being correct Bayes’s Theorem: Joint vs. Conditional Probability

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Automated Reasoning using Probability: Inference Tasks Adapted from slide © 2004 S. Russell & P. Norvig. Reused with permission.

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Bayesian Inference: Assessment Answering User Queries  Suppose we want to perform intelligent inferences over a database DB  Scenario 1: DB contains records (instances), some “labeled” with answers  Scenario 2: DB contains probabilities (annotations) over propositions  QA: an application of probabilistic inference QA Using Prior and Conditional Probabilities: Example  Query: Does patient have cancer or not?  Suppose: patient takes a lab test and result comes back positive  Correct + result in only 98% of cases in which disease is actually present  Correct - result in only 97% of cases in which disease is not present  Only of the entire population has this cancer    P(false negative for H 0  Cancer) = 0.02 (NB: for 1-point sample)    P(false positive for H 0  Cancer) = 0.03 (NB: for 1-point sample)  P(+ | H 0 ) P(H 0 ) = , P(+ | H A ) P(H A ) =  h MAP = H A   Cancer

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Bayes’s Theorem MAP Hypothesis  Generally want most probable hypothesis given training data  Define:  value of x in sample space  with highest f(x)  Maximum a posteriori hypothesis, h MAP ML Hypothesis  Assume that p(h i ) = p(h j ) for all pairs i, j (uniform priors, i.e., P H ~ Uniform)  Can further simplify and choose maximum likelihood hypothesis, h ML Choosing Hypotheses

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Graphical Models of Probability P(20s, Female, Low, Non-Smoker, No-Cancer, Negative, Negative) = P(T) · P(F) · P(L | T) · P(N | T, F) · P(N | L, N) · P(N | N) · P(N | N) Conditional Independence  X is conditionally independent (CI) from Y given Z iff P(X | Y, Z) = P(X | Z) for all values of X, Y, and Z  Example: P(Thunder | Rain, Lightning) = P(Thunder | Lightning)  T  R | L Bayesian (Belief) Network  Acyclic directed graph model B = (V, E,  ) representing CI assertions over   Vertices (nodes) V: denote events (each a random variable)  Edges (arcs, links) E: denote conditional dependencies Markov Condition for BBNs (Chain Rule): Example BBN X1X1 X3X3 X4X4 X5X5 Age Exposure-To-Toxins Smoking Cancer X6X6 Serum Calcium X2X2 Gender X7X7 Lung Tumor

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Burglar Network (Pearl, 1986) aka Burglar Alarm Network BNJ 3 © 2004 Kansas State University Burglar aka Burglar Alarm Network Pearl (1986)

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Semantics of Bayesian Networks Adapted from slide © 2004 S. Russell & P. Norvig. Reused with permission.

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Markov Blanket Adapted from slide © 2004 S. Russell & P. Norvig. Reused with permission.

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Constructing Bayesian Networks: Chain Rule in Inference & Learning Adapted from slide © 2004 S. Russell & P. Norvig. Reused with permission.

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Evidential Reasoning: Example – Car Diagnosis Adapted from slide © 2004 S. Russell & P. Norvig. Reused with permission.

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Asia Network aka Chest Clinic Network BNJ 3 © 2004 Kansas State University Asia aka Chest Clinic Network

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence ALARM: A Logical Alarm Reduction Mechanism BNJ 3 © 2004 Kansas State University ALARM Network © 1990 Cooper et al.

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Tools for Building Graphical Models Commercial Tools: Ergo, Netica, TETRAD, Hugin Bayes Net Toolbox (BNT) – Murphy (1997-present)  Distribution pagehttp://http.cs.berkeley.edu/~murphyk/Bayes/bnt.htmlhttp://http.cs.berkeley.edu/~murphyk/Bayes/bnt.html  Development grouphttp://groups.yahoo.com/group/BayesNetToolboxhttp://groups.yahoo.com/group/BayesNetToolbox Bayesian Network tools in Java (BNJ) – Hsu et al. (1999-present)  Distribution pagehttp://bnj.sourceforge.nethttp://bnj.sourceforge.net  Development grouphttp://groups.yahoo.com/group/bndevhttp://groups.yahoo.com/group/bndev  Current (re)implementation projects for KSU KDD Lab Continuous state: Minka (2002) – Hsu, Guo, Li Formats: XML BNIF (MSBN), Netica – Barber, Guo Space-efficient DBN inference – Meyer Bounded cutset conditioning – Chandak

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence References: Graphical Models & Inference Graphical Models  Bayesian (Belief) Networks tutorial – Murphy (2001)  Learning Bayesian Networks – Heckerman (1996, 1999) Inference Algorithms  Junction Tree (Join Tree, L-S, Hugin): Lauritzen & Spiegelhalter (1988)  (Bounded) Loop Cutset Conditioning: Horvitz & Cooper (1989)  Variable Elimination (Bucket Elimination, ElimBel): Dechter (1986)  Recommended Books Neapolitan (1990) – out of print; see Pearl (1988), Jensen (2001) Castillo, Gutierrez, Hadi (1997) Cowell, Dawid, Lauritzen, Spiegelhalter (1999)  Stochastic Approximation

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Terminology Uncertain Reasoning: Inference Task with Uncertain Premises, Rules Probabilistic Representation  Views of probability  Subjectivist: measure of belief in sentences  Frequentist: likelihood ratio  Logicist: counting evidence  Founded on Kolmogorov axioms  Sum rule  Prior, joint vs. conditional  Bayes’s theorem & product rule: P(A | B) = (P(B | A) * P(A)) / P(B)  Independence & conditional independence Probabilistic Reasoning  Inference by enumeration  Evidential reasoning

Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Last Class: Reasoning under Uncertainty and Probability (Ch. 13)  Uncertainty is pervasive  What are we uncertain about? Today: Chapter 13 Concluded, Preview of Chapter 14  Why probability  Axiomatic basis: Kolmogorov  With utility theory: sound foundation of rational decision making  Joint probability  Independence  Probabilistic reasoning: inference by enumeration  Conditioning  Bayes’s theorem (aka Bayes’ rule)  Conditional independence Coming Week: More Applied Probability  Graphical models as KR for uncertainty: Bayesian networks, etc.  Some inference algorithms for Bayes nets Summary Points