Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)

Slides:



Advertisements
Similar presentations
CS188: Computational Models of Human Behavior
Advertisements

Bayesian networks Chapter 14 Section 1 – 2. Outline Syntax Semantics Exact computation.
Bayesian Networks CSE 473. © Daniel S. Weld 2 Last Time Basic notions Atomic events Probabilities Joint distribution Inference by enumeration Independence.
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
Uncertainty Everyday reasoning and decision making is based on uncertain evidence and inferences. Classical logic only allows conclusions to be strictly.
For Monday Read chapter 18, sections 1-2 Homework: –Chapter 14, exercise 8 a-d.
Homework 3: Naive Bayes Classification
For Monday Finish chapter 14 Homework: –Chapter 13, exercises 8, 15.
CPSC 322, Lecture 26Slide 1 Reasoning Under Uncertainty: Belief Networks Computer Science cpsc322, Lecture 27 (Textbook Chpt 6.3) March, 16, 2009.
Bayesian network inference
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11
Probabilistic Reasoning Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 14 (14.1, 14.2, 14.3, 14.4) Capturing uncertain knowledge Probabilistic.
Bayesian networks Chapter 14 Section 1 – 2.
Bayes Nets Rong Jin. Hidden Markov Model  Inferring from observations (o i ) to hidden variables (q i )  This is a general framework for representing.
Bayesian Belief Networks
Bayesian Networks. Graphical Models Bayesian networks Conditional random fields etc.
Graphical Models Lei Tang. Review of Graphical Models Directed Graph (DAG, Bayesian Network, Belief Network) Typically used to represent causal relationship.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
Bayes Nets. Bayes Nets Quick Intro Topic of much current research Models dependence/independence in probability distributions Graph based - aka “graphical.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
1 A Tutorial on Bayesian Networks Modified by Paul Anderson from slides by Weng-Keen Wong School of Electrical Engineering and Computer Science Oregon.
Probabilistic Reasoning
Bayesian Networks Textbook: Probabilistic Reasoning, Sections 1-2, pp
Bayesian networks Chapter 14. Outline Syntax Semantics.
Bayesian networks Chapter 14 Section 1 – 2. Bayesian networks A simple, graphical notation for conditional independence assertions and hence for compact.
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Digital Statisticians INST 4200 David J Stucki Spring 2015.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
For Wednesday Read Chapter 11, sections 1-2 Program 2 due.
Baye’s Rule.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
Bayesian Statistics and Belief Networks. Overview Book: Ch 13,14 Refresher on Probability Bayesian classifiers Belief Networks / Bayesian Networks.
1 Monte Carlo Artificial Intelligence: Bayesian Networks.
Introduction to Bayesian Networks
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
CSE PR 1 Reasoning - Rule-based and Probabilistic Representing relations with predicate logic Limitations of predicate logic Representing relations.
Bayesian approaches to knowledge representation and reasoning Part 1 (Chapter 13)
Bayesian Nets and Applications. Naïve Bayes 2  What happens if we have more than one piece of evidence?  If we can assume conditional independence 
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
Marginalization & Conditioning Marginalization (summing out): for any sets of variables Y and Z: Conditioning(variant of marginalization):
1 BN Semantics 1 Graphical Models – Carlos Guestrin Carnegie Mellon University September 15 th, 2008 Readings: K&F: 3.1, 3.2, –  Carlos.
CHAPTER 5 Probability Theory (continued) Introduction to Bayesian Networks.
1 Bayesian Networks: A Tutorial. 2 Introduction Suppose you are trying to determine if a patient has tuberculosis. You observe the following symptoms:
Quiz 3: Mean: 9.2 Median: 9.75 Go over problem 1.
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Reasoning Under Uncertainty: Independence and Inference CPSC 322 – Uncertainty 5 Textbook §6.3.1 (and for HMMs) March 25, 2011.
Introduction on Graphic Models
Belief Networks Kostas Kontogiannis E&CE 457. Belief Networks A belief network is a graph in which the following holds: –A set of random variables makes.
Conditional Independence As with absolute independence, the equivalent forms of X and Y being conditionally independent given Z can also be used: P(X|Y,
Probabilistic Reasoning Inference and Relational Bayesian Networks.
Weng-Keen Wong, Oregon State University © Bayesian Networks: A Tutorial Weng-Keen Wong School of Electrical Engineering and Computer Science Oregon.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
Reasoning Under Uncertainty: More on BNets structure and construction
Bayesian networks Chapter 14 Section 1 – 2.
Inference in Bayesian Networks
Reasoning Under Uncertainty: More on BNets structure and construction
Reasoning Under Uncertainty: Conditioning, Bayes Rule & Chain Rule
Learning Bayesian Network Models from Data
Uncertainty in AI.
Structure and Semantics of BN
CAP 5636 – Advanced Artificial Intelligence
CS 188: Artificial Intelligence
Class #16 – Tuesday, October 26
CS 188: Artificial Intelligence Spring 2007
Structure and Semantics of BN
Chapter 14 February 26, 2004.
CS 188: Artificial Intelligence Fall 2008
Presentation transcript:

Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)

Reading for this week S. Wooldridge, Bayesian belief networks (linked from class website) Textbook, Chapter 9, Unsupervised Learning (skip section 9.3)Bayesian belief networks

Bayesian Networks

A patient comes into a doctor’s office with a fever and a bad cough. Hypothesis space H: h 1 : patient has flu h 2 : patient does not have flu Data D: coughing = true, fever = true,, smokes = true

Naive Bayes smokesflucoughfever Cause Effects

In principle, the full joint distribution can be used to answer any question about probabilities of these combined parameters. However, size of full joint distribution scales exponentially with number of parameters so is expensive to store and to compute with. Full joint probability distribution Fever  Fever Fever  Fever flu p 1 p 2 p 3 p 4  flu p 5 p 6 p 7 p 8 cough  cough Sum of all boxes is 1. fever  fever fever  fever flu p 9 p 10 p 11 p 12  flu p 13 p 14 p 15 p 16 cough  cough  smokes smokes

For example, what if we had another attribute, “allergies”? How many probabilities would we need to specify? Full joint probability distribution Fever  Fever Fever  Fever flu p 1 p 2 p 3 p 4  flu p 5 p 6 p 7 p 8 cough  cough fever  fever fever  fever flu p 9 p 10 p 11 p 12  flu p 13 p 14 p 15 p 16 cough  cough  smokes smokes

Fever  Fever Fever  Fever flu p 1 p 2 p 3 p 4  flu p 5 p 6 p 7 p 8 cough  cough fever  fever fever  fever flu p 9 p 10 p 11 p 12  flu p 13 p 14 p 15 p 16 cough  cough  smokes smokes Allergy Fever  Fever Fever  Fever flu p 17 p 18 p 19 p 20  flu p 21 p 22 p 23 p 24 cough  cough fever  fever fever  fever flu p 25 p 26 p 27 p 28  flu p 29 p 30 p 31 p 32 cough  cough  smokes smokes  Allergy

Fever  Fever Fever  Fever flu p 1 p 2 p 3 p 4  flu p 5 p 6 p 7 p 8 cough  cough fever  fever fever  fever flu p 9 p 10 p 11 p 12  flu p 13 p 14 p 15 p 16 cough  cough  smokes smokes Allergy Fever  Fever Fever  Fever flu p 17 p 18 p 19 p 20  flu p 21 p 22 p 23 p 24 cough  cough fever  fever fever  fever flu p 25 p 26 p 27 p 28  flu p 29 p 30 p 31 p 32 cough  cough  smokes smokes  Allergy But can reduce this if we know which variables are conditionally independent

Bayesian networks Idea is to represent dependencies (or causal relations) for all the variables so that space and computation-time requirements are minimized. smokes flu cough fever “Graphical Models”

Bayesian Networks = Bayesian Belief Networks = Bayes Nets Bayesian Network: Alternative representation for complete joint probability distribution “Useful for making probabilistic inference about models domains characterized by inherent complexity and uncertainty” Uncertainty can come from: – incomplete knowledge of domain – inherent randomness in behavior in domain

true0.01 false0.99 flu smoke cough fever truefalse true false fever flu true0.2 false0.8 smoke truefalse True TrueFalse FalseTrue false cough smokeflu Conditional probability tables for each node Example:

Inference in Bayesian networks If network is correct, can calculate full joint probability distribution from network. where parents(X i ) denotes specific values of parents of X i.

Example Calculate

In general... If network is correct, can calculate full joint probability distribution from network. where parents(X i ) denotes specific values of parents of X i. But need efficient algorithms to do this (e.g., “belief propagation”, “Markov Chain Monte Carlo”).

Example from the reading:

What is the probability that Student A is late? What is the probability that Student B is late?

What is the probability that Student A is late? What is the probability that Student B is late? Unconditional (“marginal”) probability. We don’t know if there is a train strike.

Now, suppose we know that there is a train strike. How does this revise the probability that the students are late?

Evidence: There is a train strike.

Now, suppose we know that Student A is late. How does this revise the probability that there is a train strike? How does this revise the probability that Student B is late? Notion of “belief propagation”. Evidence: Student A is late.

Now, suppose we know that Student A is late. How does this revise the probability that there is a train strike? How does this revise the probability that Student B is late? Notion of “belief propagation”. Evidence: Student A is late.

Another example from the reading: pneumoniasmoking temperaturecough

Three types of inference Diagnostic: Use evidence of an effect to infer probability of a cause. – E.g., Evidence: cough=true. What is P(pneumonia | cough)? Causal inference: Use evidence of a cause to infer probability of an effect – E.g., Evidence: pneumonia=true. What is P(cough | pneumonia)? Inter-causal inference: “Explain away” potentially competing causes of a shared effect. – E.g., Evidence: smoking=true. What is P(pneumonia | cough and smoking)?

Diagnostic: Evidence: cough=true. What is P(pneumonia | cough)? pneumoniasmoking temperaturecough

Diagnostic: Evidence: cough=true. What is P(pneumonia | cough)? pneumoniasmoking temperaturecough

Causal: Evidence: pneumonia=true. What is P(cough | pneumonia)? pneumoniasmoking temperaturecough

Causal: Evidence: pneumonia=true. What is P(cough | pneumonia)? pneumoniasmoking temperaturecough

Inter-causal: Evidence: smoking=true. What is P(pneumonia | cough and smoking)? pneumoniasmoking temperaturecough

Math we used: Definition of conditional probability: Bayes Theorem Unconditional (marginal) probability Probability inference in Bayesian networks:

Complexity of Bayesian Networks For n random Boolean variables: Full joint probability distribution: 2 n entries Bayesian network with at most k parents per node: – Each conditional probability table: at most 2 k entries – Entire network: n 2 k entries

What are the advantages of Bayesian networks? Intuitive, concise representation of joint probability distribution (i.e., conditional dependencies) of a set of random variables. Represents “beliefs and knowledge” about a particular class of situations. Efficient (approximate) inference algorithms Efficient, effective learning algorithms

Issues in Bayesian Networks Building / learning network topology Assigning / learning conditional probability tables Approximate inference via sampling

Real-World Example: The Lumière Project at Microsoft Research Bayesian network approach to answering user queries about Microsoft Office. “At the time we initiated our project in Bayesian information retrieval, managers in the Office division were finding that users were having difficulty finding assistance efficiently.” “As an example, users working with the Excel spreadsheet might have required assistance with formatting “a graph”. Unfortunately, Excel has no knowledge about the common term, “graph,” and only considered in its keyword indexing the term “chart”.

Networks were developed by experts from user modeling studies.

Offspring of project was Office Assistant in Office 97, otherwise known as “clippie”.