Lecture 2: Bayesian Decision Theory 1. Diagram and formulation

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

ECE 8443 – Pattern Recognition Objectives: Course Introduction Typical Applications Resources: Syllabus Internet Books and Notes D.H.S: Chapter 1 Glossary.
Pattern Classification. Chapter 2 (Part 1): Bayesian Decision Theory (Sections ) Introduction Bayesian Decision Theory–Continuous Features.
Bayesian Decision Theory
Bayesian Decision Theory Chapter 2 (Duda et al.) – Sections
Visual Recognition Tutorial1 Bayesian decision making with discrete probabilities – an example Looking at continuous densities Bayesian decision.
Chapter 2: Bayesian Decision Theory (Part 1) Introduction Bayesian Decision Theory–Continuous Features All materials used in this course were taken from.
Machine Learning CMPT 726 Simon Fraser University
Pattern Classification, Chapter 1 1 Basic Probability.
Introduction to Bayesian Learning Ata Kaban School of Computer Science University of Birmingham.
Thanks to Nir Friedman, HU
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
1. Introduction to Pattern Recognition and Machine Learning. Prof. A.L. Yuille. Dept. Statistics. UCLA. Stat 231. Fall 2004.
Lecture 1: Introduction to Pattern Recognition
Bayesian Decision Theory Making Decisions Under uncertainty 1.
Institute of Systems and Robotics ISR – Coimbra Mobile Robotics Lab Bayesian Approaches 1 1 jrett.
Principles of Pattern Recognition
Speech Recognition Pattern Classification. 22 September 2015Veton Këpuska2 Pattern Classification  Introduction  Parametric classifiers  Semi-parametric.
Classification. An Example (from Pattern Classification by Duda & Hart & Stork – Second Edition, 2001)
IBS-09-SL RM 501 – Ranjit Goswami 1 Basic Probability.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 02: BAYESIAN DECISION THEORY Objectives: Bayes.
2. Bayes Decision Theory Prof. A.L. Yuille Stat 231. Fall 2004.
Sergios Theodoridis Konstantinos Koutroumbas Version 2
Empirical Research Methods in Computer Science Lecture 7 November 30, 2005 Noah Smith.
URL:.../publications/courses/ece_8443/lectures/current/exam/2004/ ECE 8443 – Pattern Recognition LECTURE 15: EXAM NO. 1 (CHAP. 2) Spring 2004 Solutions:
1 E. Fatemizadeh Statistical Pattern Recognition.
Optimal Bayes Classification
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Lecture notes 9 Bayesian Belief Networks.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 8 Sept 23, 2005 Nanjing University of Science & Technology.
1 Bayesian Decision Theory Shyh-Kang Jeng Department of Electrical Engineering/ Graduate Institute of Communication/ Graduate Institute of Networking and.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
ECE 471/571 – Lecture 2 Bayesian Decision Theory 08/25/15.
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
Bayesian Decision Theory Basic Concepts Discriminant Functions The Normal Density ROC Curves.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 04: GAUSSIAN CLASSIFIERS Objectives: Whitening.
Lecture 3: MLE, Bayes Learning, and Maximum Entropy
Intro. ANN & Fuzzy Systems Lecture 15. Pattern Classification (I): Statistical Formulation.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 9: Review.
Basic Technical Concepts in Machine Learning Introduction Supervised learning Problems in supervised learning Bayesian decision theory.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Bayes Rule Mutual Information Conditional.
8/16/99 Computer Vision: Vision and Modeling. 8/16/99 Lucas-Kanade Extensions Support Maps / Layers: Robust Norm, Layered Motion, Background Subtraction,
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
1 1)Bayes’ Theorem 2)MAP, ML Hypothesis 3)Bayes optimal & Naïve Bayes classifiers IES 511 Machine Learning Dr. Türker İnce (Lecture notes by Prof. T. M.
Objectives: Loss Functions Risk Min. Error Rate Class. Resources: DHS – Chap. 2 (Part 1) DHS – Chap. 2 (Part 2) RGO - Intro to PR MCE for Speech MCE for.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Lecture 2. Bayesian Decision Theory
Basic Technical Concepts in Machine Learning
Lecture 1.31 Criteria for optimal reception of radio signals.
Special Topics In Scientific Computing
LECTURE 03: DECISION SURFACES
Comp328 tutorial 3 Kai Zhang
Special Topics In Scientific Computing
Data Mining Lecture 11.
Lecture 26: Faces and probabilities
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
CSSE463: Image Recognition Day 27
Advanced Pattern Recognition
An Introduction to Supervised Learning
Pattern Recognition and Image Analysis
Advanced Pattern Recognition
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Recognition and Machine Learning
CSSE463: Image Recognition Day 31
LECTURE 23: INFORMATION THEORY REVIEW
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Machine Learning: Lecture 6
Bayesian Decision Theory
Presentation transcript:

Lecture 2: Bayesian Decision Theory 1. Diagram and formulation 2. Bayes rule for inference 3. Bayesian decision 4. Discriminant functions and space partition 5. Advanced issues Lecture note for Stat 231: Pattern Recognition and Machine Learning

Diagram of pattern classification Procedure of pattern recognition and decision making Observables X Features x Inner belief w Action a subjects X--- all the observables using existing sensors and instruments x --- is a set of features selected from components of X, or linear/non-linear functions of X. w --- is our inner belief/perception about the subject class. --- is the action that we take for x. We denote the three spaces by Lecture note for Stat 231: Pattern Recognition and Machine Learning

Examples Ex 1: Fish classification X=I is the image of fish, x =(brightness, length, fin#, ….) w is our belief what the fish type is Wc={“sea bass”, “salmon”, “trout”, …} a is a decision for the fish type, in this case Wc= Wa Wa ={“sea bass”, “salmon”, “trout”, …} Ex 2: Medical diagnosis X= all the available medical tests, imaging scans that a doctor can order for a patient x =(blood pressure, glucose level, cough, x-ray….) w is an illness type Wc={“Flu”, “cold”, “TB”, “pneumonia”, “lung cancer”…} a is a decision for treatment, Wa ={“Tylenol”, “Hospitalize”, …} Lecture note for Stat 231: Pattern Recognition and Machine Learning

Tasks Observables X Features x Inner belief w Decision a subjects control sensors selecting Informative features statistical inference risk/cost minimization In Bayesian decision theory, we are concerned with the last three steps in the big ellipse assuming that the observables are given and features are selected. Lecture note for Stat 231: Pattern Recognition and Machine Learning

Bayesian Decision Theory statistical Inference risk/cost minimization Features x Inner belief p(w|x) Decision a(x) Two probability tables: a). Prior p(w) b). Likelihood p(x|w) A risk/cost function (is a two-way table) l(a | w) The belief on the class w is computed by the Bayes rule The risk is computed by Lecture note for Stat 231: Pattern Recognition and Machine Learning

Decision Rule A decision rule is a mapping function from feature space to the set of actions we will show that randomized decisions won’t be optimal. A decision is made to minimize the average cost / risk, It is minimized when our decision is made to minimize the cost / risk for each instance x. Lecture note for Stat 231: Pattern Recognition and Machine Learning

Bayesian error In a special case, like fish classification, the action is classification, we assume a 0/1 error. The risk for classifying x to class ai is, The optimal decision is to choose the class that has maximum posterior probability The total risk for a decision rule, in this case, is called the Bayesian error Lecture note for Stat 231: Pattern Recognition and Machine Learning

Discriminant functions To summarize, we take an action to maximize some discriminant functions Lecture note for Stat 231: Pattern Recognition and Machine Learning

Partition of feature space The decision is a partition /coloring of the feature space into k subspaces 5 3 2 1 4 Lecture note for Stat 231: Pattern Recognition and Machine Learning

An example of fish classification Lecture note for Stat 231: Pattern Recognition and Machine Learning

Decision/classification Boundaries Lecture note for Stat 231: Pattern Recognition and Machine Learning

Close-form solutions In two case classification, k=2, with p(x|w) being Normal densities, the decision Boundaries can be computed in close-form. Lecture note for Stat 231: Pattern Recognition and Machine Learning

Advanced issues Subjectivism of the prior in Bayesian decision Learning the probabilities p(w) and p(x|w) ---machine learning Choosing optimal features---what features are most discriminative? 4. How many features are enough? 5. Consider the context or sequential information in classification Lecture note for Stat 231: Pattern Recognition and Machine Learning