Does Naïve Bayes always work?

Slides:



Advertisements
Similar presentations
BAYESIAN NETWORKS Ivan Bratko Faculty of Computer and Information Sc. University of Ljubljana.
Advertisements

Classification. Introduction A discriminant is a function that separates the examples of different classes. For example – IF (income > Q1 and saving >Q2)
Representations for KBS: Uncertainty & Decision Support
Rulebase Expert System and Uncertainty. Rule-based ES Rules as a knowledge representation technique Type of rules :- relation, recommendation, directive,
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
Probability: Review The state of the world is described using random variables Probabilities are defined over events –Sets of world states characterized.
PROBABILITY. Uncertainty  Let action A t = leave for airport t minutes before flight from Logan Airport  Will A t get me there on time ? Problems :
FT228/4 Knowledge Based Decision Support Systems
Knowledge Representation and Reasoning
5/17/20151 Probabilistic Reasoning CIS 479/579 Bruce R. Maxim UM-Dearborn.
Introduction of Probabilistic Reasoning and Bayesian Networks
1 Slides for the book: Probabilistic Robotics Authors: Sebastian Thrun Wolfram Burgard Dieter Fox Publisher: MIT Press, Web site for the book & more.
AI – CS364 Uncertainty Management Introduction to Uncertainty Management 21 st September 2006 Dr Bogdan L. Vrusias
1 Abduction, Uncertainty, and Probabilistic Reasoning Yun Peng UMBC.
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
Probabilities Random Number Generators –Actually pseudo-random –Seed Same sequence from same seed Often time is used. Many examples on web. Custom random.
1.Examples of using probabilistic ideas in robotics 2.Reverend Bayes and review of probabilistic ideas 3.Introduction to Bayesian AI 4.Simple example.
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
Bayesian Belief Networks
Bayesian Networks. Graphical Models Bayesian networks Conditional random fields etc.
1 Bayesian Reasoning Chapter 13 CMSC 471 Adapted from slides by Tim Finin and Marie desJardins.
Naïve Bayes Model. Outline Independence and Conditional Independence Naïve Bayes Model Application: Spam Detection.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
1 CMSC 471 Fall 2002 Class #19 – Monday, November 4.
1 NA387 Lecture 6: Bayes’ Theorem, Independence Devore, Sections: 2.4 – 2.5.
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
Naive Bayes Classifier
1 Abduction, Uncertainty, and Probabilistic Reasoning Chapters 13, 14, and more.
1 Robot Environment Interaction Environment perception provides information about the environment’s state, and it tends to increase the robot’s knowledge.
1 Abduction, Uncertainty, and Probabilistic Reasoning Chapter 15 and more.
Introduction to Bayesian Networks
Bayesian Classification. Bayesian Classification: Why? A statistical classifier: performs probabilistic prediction, i.e., predicts class membership probabilities.
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 6.1: Bayes Filter Jürgen Sturm Technische Universität München.
Uncertainty Management in Rule-based Expert Systems
Uncertainty in Expert Systems
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Tractable Inference for Complex Stochastic Processes X. Boyen & D. Koller Presented by Shiau Hong Lim Partially based on slides by Boyen & Koller at UAI.
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
Multi-target Detection in Sensor Networks Xiaoling Wang ECE691, Fall 2003.
Textbook Basics of an Expert System: – “Expert systems: Design and Development,” by: John Durkin, 1994, Chapters 1-4. Uncertainty (Probability, Certainty.
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
CSE 473 Uncertainty. © UW CSE AI Faculty 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one has stood.
CHAPTER 3: BAYESIAN DECISION THEORY. Making Decision Under Uncertainty Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Naïve Bayes Classifier April 25 th, Classification Methods (1) Manual classification Used by Yahoo!, Looksmart, about.com, ODP Very accurate when.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Probabilistic Robotics Probability Theory Basics Error Propagation Slides from Autonomous Robots (Siegwart and Nourbaksh), Chapter 5 Probabilistic Robotics.
CSE (c) S. Tanimoto, 2007 Bayes Nets 1 Bayes Networks Outline: Why Bayes Nets? Review of Bayes’ Rule Combining independent items of evidence General.
Matching ® ® ® Global Map Local Map … … … obstacle Where am I on the global map?                                   
Naive Bayes Classifier. REVIEW: Bayesian Methods Our focus this lecture: – Learning and classification methods based on probability theory. Bayes theorem.
Bayesian Reasoning Chapter 13 Thomas Bayes,
Bayesian Reasoning Chapter 13 Thomas Bayes,
Review of Probability.
Does Naïve Bayes always work?
Chapter 4 Probability.
Naive Bayes Classifier
Today.
Markov ó Kalman Filter Localization
CSE-490DF Robotics Capstone
Professor Marie desJardins,
Class #21 – Monday, November 10
Bayesian Reasoning Chapter 13 Thomas Bayes,
Bayesian Reasoning Chapter 13 Thomas Bayes,
CS 188: Artificial Intelligence Fall 2007
Chapter 14 February 26, 2004.
Presentation transcript:

Does Naïve Bayes always work? No

Bayes’ Theorem with relative likelihood In the setting of diagnostic/evidential reasoning Known prior probability of hypothesis conditional probability Want to compute the posterior probability Bayes’ theorem (formula 1): If the purpose is to find which of the n hypotheses is more plausible given , then we can ignore the denominator and rank them, use relative likelihood

Relative likelihood can be computed from and , if we assume all hypotheses are ME and EXH Then we have another version of Bayes’ theorem: where , the sum of relative likelihood of all n hypotheses, is a normalization factor

Naïve Bayesian Approach Knowledge base: Case input: Find the hypothesis with the highest posterior probability By Bayes’ theorem Assume all pieces of evidence are conditionally independent, given any hypothesis

absolute posterior probability The relative likelihood The absolute posterior probability Evidence accumulation (when new evidence is discovered)

Discussion of Assumptions of Naïve Bayes Assumption 1: hypotheses are mutually exclusive and exhaustive Single fault assumption (one and only one hypothesis must be true) Multi-faults do exist in individual cases Can be viewed as an approximation of situations where hypotheses are independent of each other and their prior probabilities are very small Assumption 2: pieces of evidence are conditionally independent of each other, given any hypothesis Manifestations themselves are not independent of each other, they are correlated by their common causes Reasonable under single fault assumption Not so when multi-faults are to be considered

Limitations of the naïve Bayesian system

Limitations of the naïve Bayesian system Cannot handle hypotheses of multiple disorders well Suppose are independent of each other Consider a composite hypothesis How to compute the posterior probability (or relative likelihood) Using Bayes’ theorem

P(B|A, E) <<P(B|A) but this is a very unreasonable assumption Need a better representation and a better assumption Earthquake? Burgler? Alarm is on E and B are independent But when A is given, they are (adversely) dependent because they become competitors to explain A P(B|A, E) <<P(B|A) E explains away of A E: earth quake B: burglar A: alarm set off

Naïve Bayes cannot handle causal chaining Example. A: weather of the year B: cotton production of the year C: cotton price of next year Observed: A influences C The influence is not direct (A -> B -> C) P(C|B, A) = P(C|B): instantiation of B blocks influence of A on C Need a better representation and a better assumption A B C

Bayesian Networks and Markov Models – applications in robotics Bayesian AI Bayesian Filters Kalman Filters Particle Filters Bayesian networks Decision networks Reasoning about changes over time Dynamic Bayesian Networks Markov models