Does Naïve Bayes always work?

Slides:



Advertisements
Similar presentations
Motivating Markov Chain Monte Carlo for Multiple Target Tracking
Advertisements

BAYESIAN NETWORKS Ivan Bratko Faculty of Computer and Information Sc. University of Ljubljana.
State Estimation and Kalman Filtering CS B659 Spring 2013 Kris Hauser.
Classification. Introduction A discriminant is a function that separates the examples of different classes. For example – IF (income > Q1 and saving >Q2)
Representations for KBS: Uncertainty & Decision Support
Bayesian Learning Provides practical learning algorithms
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
Probability: Review The state of the world is described using random variables Probabilities are defined over events –Sets of world states characterized.
PROBABILITY. Uncertainty  Let action A t = leave for airport t minutes before flight from Logan Airport  Will A t get me there on time ? Problems :
FT228/4 Knowledge Based Decision Support Systems
Knowledge Representation and Reasoning
5/17/20151 Probabilistic Reasoning CIS 479/579 Bruce R. Maxim UM-Dearborn.
1 Slides for the book: Probabilistic Robotics Authors: Sebastian Thrun Wolfram Burgard Dieter Fox Publisher: MIT Press, Web site for the book & more.
AI – CS364 Uncertainty Management Introduction to Uncertainty Management 21 st September 2006 Dr Bogdan L. Vrusias
1 Abduction, Uncertainty, and Probabilistic Reasoning Yun Peng UMBC.
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
Review: Bayesian learning and inference
1.Examples of using probabilistic ideas in robotics 2.Reverend Bayes and review of probabilistic ideas 3.Introduction to Bayesian AI 4.Simple example.
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
Bayesian Belief Networks
1 Bayesian Reasoning Chapter 13 CMSC 471 Adapted from slides by Tim Finin and Marie desJardins.
Naïve Bayes Model. Outline Independence and Conditional Independence Naïve Bayes Model Application: Spam Detection.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
1 CMSC 471 Fall 2002 Class #19 – Monday, November 4.
Does Naïve Bayes always work?
Markov Localization & Bayes Filtering
1 NA387 Lecture 6: Bayes’ Theorem, Independence Devore, Sections: 2.4 – 2.5.
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
Naive Bayes Classifier
1 Robot Environment Interaction Environment perception provides information about the environment’s state, and it tends to increase the robot’s knowledge.
1 Abduction, Uncertainty, and Probabilistic Reasoning Chapter 15 and more.
Introduction to Bayesian Networks
Bayesian Classification. Bayesian Classification: Why? A statistical classifier: performs probabilistic prediction, i.e., predicts class membership probabilities.
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 6.1: Bayes Filter Jürgen Sturm Technische Universität München.
Uncertainty Management in Rule-based Expert Systems
Uncertainty in Expert Systems
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Tractable Inference for Complex Stochastic Processes X. Boyen & D. Koller Presented by Shiau Hong Lim Partially based on slides by Boyen & Koller at UAI.
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
Multi-target Detection in Sensor Networks Xiaoling Wang ECE691, Fall 2003.
Textbook Basics of an Expert System: – “Expert systems: Design and Development,” by: John Durkin, 1994, Chapters 1-4. Uncertainty (Probability, Certainty.
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
CSE 473 Uncertainty. © UW CSE AI Faculty 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one has stood.
CHAPTER 3: BAYESIAN DECISION THEORY. Making Decision Under Uncertainty Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Naïve Bayes Classifier April 25 th, Classification Methods (1) Manual classification Used by Yahoo!, Looksmart, about.com, ODP Very accurate when.
Probabilistic Robotics Probability Theory Basics Error Propagation Slides from Autonomous Robots (Siegwart and Nourbaksh), Chapter 5 Probabilistic Robotics.
Naive Bayes Classifier. REVIEW: Bayesian Methods Our focus this lecture: – Learning and classification methods based on probability theory. Bayes theorem.
Bayesian Reasoning Chapter 13 Thomas Bayes,
Bayesian Reasoning Chapter 13 Thomas Bayes,
Review of Probability.
CS 2750: Machine Learning Directed Graphical Models
Bayesian and Markov Test
Chapter 4 Probability.
Naive Bayes Classifier
Today.
Markov ó Kalman Filter Localization
State Estimation Probability, Bayes Filtering
CSE-490DF Robotics Capstone
Professor Marie desJardins,
Class #21 – Monday, November 10
Bayesian Reasoning Chapter 13 Thomas Bayes,
Bayesian Reasoning Chapter 13 Thomas Bayes,
CS 188: Artificial Intelligence Fall 2007
28th September 2005 Dr Bogdan L. Vrusias
DR.M.THIAGARAJAN ASSOCIATE PROFESSOR OF MATHEMATICS
Presentation transcript:

Does Naïve Bayes always work? No

Bayes’ Theorem with relative likelihood In the setting of diagnostic/evidential reasoning Known prior probability of hypothesis conditional probability Want to compute the posterior probability Bayes’ theorem (formula 1): If the purpose is to find which of the n hypotheses is more plausible given , then we can ignore the denominator and rank them, use relative likelihood

Relative likelihood can be computed from and , if we assume all hypotheses are ME and EXH Then we have another version of Bayes’ theorem: where , the sum of relative likelihood of all n hypotheses, is a normalization factor

Naïve Bayesian Approach Knowledge base: Case input: Find the hypothesis with the highest posterior probability By Bayes’ theorem Assume all pieces of evidence are conditionally independent, given any hypothesis

absolute posterior probability The relative likelihood The absolute posterior probability Evidence accumulation (when new evidence is discovered)

Discussion of Assumptions of Naïve Bayes Assumption 1: hypotheses are mutually exclusive and exhaustive Single fault assumption (one and only one hypothesis must be true) Multi-faults do exist in individual cases Can be viewed as an approximation of situations where hypotheses are independent of each other and their prior probabilities are very small Assumption 2: pieces of evidence are conditionally independent of each other, given any hypothesis Manifestations themselves are not independent of each other, they are correlated by their common causes Reasonable under single fault assumption Not so when multi-faults are to be considered

Limitations of the naïve Bayesian system Cannot handle hypotheses of multiple disorders well Suppose are independent of each other Consider a composite hypothesis How to compute the posterior probability (or relative likelihood) Using Bayes’ theorem

P(B|A, E) <<P(B|A) but this is a very unreasonable assumption Need a better representation and a better assumption Earthquake? Burgler? Alarm is on E and B are independent But when A is given, they are (adversely) dependent because they become competitors to explain A P(B|A, E) <<P(B|A) E explains away of A E: earth quake B: burglar A: alarm set off

Naïve Bayes cannot handle causal chaining Example. A: weather of the year B: cotton production of the year C: cotton price of next year Observed: A influences C The influence is not direct (A -> B -> C) P(C|B, A) = P(C|B): instantiation of B blocks influence of A on C Need a better representation and a better assumption A B C

Bayesian Networks and Markov Models – applications in robotics Bayesian AI Bayesian Filters Kalman Filters Particle Filters Bayesian networks Decision networks Reasoning about changes over time Dynamic Bayesian Networks Markov models