URL:.../publications/courses/ece_8443/lectures/current/exam/2004/ ECE 8443 – Pattern Recognition LECTURE 15: EXAM NO. 1 (CHAP. 2) Spring 2004 Solutions:

Slides:



Advertisements
Similar presentations
Component Analysis (Review)
Advertisements

Pattern Recognition and Machine Learning
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Chapter 2: Bayesian Decision Theory (Part 2) Minimum-Error-Rate Classification Classifiers, Discriminant Functions and Decision Surfaces The Normal Density.
Pattern Classification, Chapter 2 (Part 2) 0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R.
Pattern Classification. Chapter 2 (Part 1): Bayesian Decision Theory (Sections ) Introduction Bayesian Decision Theory–Continuous Features.
Pattern Classification, Chapter 2 (Part 2) 0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R.
Chapter 2: Bayesian Decision Theory (Part 2) Minimum-Error-Rate Classification Classifiers, Discriminant Functions and Decision Surfaces The Normal Density.
Bayesian Decision Theory
Pattern Classification Chapter 2 (Part 2)0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O.
Bayesian Decision Theory Chapter 2 (Duda et al.) – Sections
Chapter 2: Bayesian Decision Theory (Part 1) Introduction Bayesian Decision Theory–Continuous Features All materials used in this course were taken from.
Machine Learning CMPT 726 Simon Fraser University
Decision Theory Naïve Bayes ROC Curves
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
METU Informatics Institute Min 720 Pattern Classification with Bio-Medical Applications PART 2: Statistical Pattern Classification: Optimal Classification.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
1 Linear Methods for Classification Lecture Notes for CMPUT 466/551 Nilanjan Ray.
Principles of Pattern Recognition
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 13 Oct 14, 2005 Nanjing University of Science & Technology.
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Lecture 2: Bayesian Decision Theory 1. Diagram and formulation
Speech Recognition Pattern Classification. 22 September 2015Veton Këpuska2 Pattern Classification  Introduction  Parametric classifiers  Semi-parametric.
ECE 8443 – Pattern Recognition LECTURE 03: GAUSSIAN CLASSIFIERS Objectives: Normal Distributions Whitening Transformations Linear Discriminants Resources.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 02: BAYESIAN DECISION THEORY Objectives: Bayes.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 03: GAUSSIAN CLASSIFIERS Objectives: Whitening.
Sergios Theodoridis Konstantinos Koutroumbas Version 2
Discriminant Functions
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Classification Heejune Ahn SeoulTech Last updated May. 03.
1 E. Fatemizadeh Statistical Pattern Recognition.
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 12 Sept 30, 2005 Nanjing University of Science & Technology.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 8 Sept 23, 2005 Nanjing University of Science & Technology.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
ECE 471/571 – Lecture 2 Bayesian Decision Theory 08/25/15.
Covariance matrices for all of the classes are identical, But covariance matrices are arbitrary.
Bayesian Decision Theory Basic Concepts Discriminant Functions The Normal Density ROC Curves.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Elements of Pattern Recognition CNS/EE Lecture 5 M. Weber P. Perona.
ECE 8443 – Pattern Recognition LECTURE 04: PERFORMANCE BOUNDS Objectives: Typical Examples Performance Bounds ROC Curves Resources: D.H.S.: Chapter 2 (Part.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 04: GAUSSIAN CLASSIFIERS Objectives: Whitening.
Lecture 3: MLE, Bayes Learning, and Maximum Entropy
Objectives: Chernoff Bound Bhattacharyya Bound ROC Curves Discrete Features Resources: V.V. – Chernoff Bound J.G. – Bhattacharyya T.T. – ROC Curves NIST.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 10: PRINCIPAL COMPONENTS ANALYSIS Objectives:
Linear Classifier Team teaching.
Introduction to Classifiers Fujinaga. Bayes (optimal) Classifier (1) A priori probabilities: and Decision rule: given and decide if and probability of.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Objectives: Loss Functions Risk Min. Error Rate Class. Resources: DHS – Chap. 2 (Part 1) DHS – Chap. 2 (Part 2) RGO - Intro to PR MCE for Speech MCE for.
Lecture 2. Bayesian Decision Theory
Lecture 1.31 Criteria for optimal reception of radio signals.
Probability Theory and Parameter Estimation I
LECTURE 03: DECISION SURFACES
Comp328 tutorial 3 Kai Zhang
Some Rules for Expectation
LECTURE 05: THRESHOLD DECODING
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
LECTURE 05: THRESHOLD DECODING
Pattern Recognition and Image Analysis
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Recognition and Machine Learning
Mathematical Foundations of BME
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Learning From Observed Data
Mathematical Foundations of BME
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
LECTURE 05: THRESHOLD DECODING
LECTURE 11: Exam No. 1 Review
Presentation transcript:

URL:.../publications/courses/ece_8443/lectures/current/exam/2004/ ECE 8443 – Pattern Recognition LECTURE 15: EXAM NO. 1 (CHAP. 2) Spring 2004 Solutions: 1a 1b 1c 2a 2b 2c 2d 2e 2f 2g

Problem No. 1: Let for a two-category one-dimensional problem with (a)Show that the minimum probability of error is given by: where Solution: The probability of error is given by: LECTURE 15: EXAM NO. 1 (CHAP. 2) PROBLEM 1 Where R1 denotes class region and R2,. To determine R1 and R2, the decision region must be determined. Calculate the decision rule with the Likelihood Ratio : For a two–category problem place in if the likelihood ratio exceeds a threshold value. Otherwise decide ( 2 ) ( 1 )

The class conditional densities are and the Prior probabilities are For a given feature, the likelihood ratio (1) can be used to determine the decision rule:,changing signs,taking log, decide Else If, decide

Assume, then, Let Where, PROBLEM 1 LECTURE 15: EXAM NO. 1 (CHAP. 2) The decision region is half way between the two means. This result makes sense from an intuitive point of view since the likelihoods are identical and differ only in their mean value. Probability of error ( 1 ) becomes: where,

Combining these two results we get: We get a similar result if  2 <  1. Where, PROBLEM 1 LECTURE 15: EXAM NO. 1 (CHAP. 2) OK. But how to be sure if this is the minimum probability of error? P[error] in terms of the posterior P[error|x] The optimal decision rule will minimize P[error|x] for every value of x. At each feature value x, P[error|x] = P[ω2|x] when ω1 was chosen. Therefore, when we integrate over the limit, the decision rule yields a minimum P[error]. This probability of error is the Bayes Error.

(b) Use the inequality to show that P e goes to zero as goes to infinity. and as the distance between the means of the two distribution tend to infinity. PROBLEM 1 LECTURE 15: EXAM NO. 1 (CHAP. 2) Solution:

(c) How does your result for (b) change if the prior probabilities are not equal? Making, puts infinite distance between the classes. The probability of error will still tend to zero if the Prior probabilities are not equal. PROBLEM 1 LECTURE 15: EXAM NO. 1 (CHAP. 2)

PROBLEM 2 Problem No. 2: Given a two-class two-dimensional classification problem (x = {x 1,x 2 }) with the following parameters (uniform distributions): LECTURE 15: EXAM NO. 1 (CHAP. 2) a)Compute the mean and covariance (hint: plot the distributions). I. Using Independent class approach, we get joint pdf as We can obtain marginal probability distribution function as

PROBLEM 2 LECTURE 15: EXAM NO. 1 (CHAP. 2) II. Using class-dependent approach, We can calculate mean and covariance as,

PROBLEM 2 LECTURE 15: EXAM NO. 1 (CHAP. 2) Figure showing the decision surface and the distribution of two classes

PROBLEM 2 b) Find the discriminant functions(e.g., g i (x) ). There are infinite numbers of solutions for the case of P(w 1 )=P(w 2 ) and P(x/w 1 )=P(x/w 2 ) in the overlap area. The simplest one can be defined as g(x)= x 1 + x 2 -1/4 such that if g(x)>0 decide class w 2 else class w 1 C) Write the Bayes decision rule for this case (hint: draw the decision boundary). Is this solution unique? Explain. Since P(w 1 )=P(w 2 ) and P(x/w 1 )=P(x/w 2 ) in the overlap area, the posterior probability will be same in the overlap area. Hence, the solution won’t be unique. d) Compute the probability of error. LECTURE 15: EXAM NO. 1 (CHAP. 2)

PROBLEM 2 e) How will the decision surface change if the priors are not equal? Explain. When the priors are equal, the decision region is at the point where two distribution meet if the distributions are similar. If the priors are not equal the decision region moves away from the higher prior class. f) How will the probability of error change if the priors are not equal? As the prior probability changes, the decision surface changes. Hence the probability of error changes. For ex, if P(w 1 )=1 and P(w 2 )=0. In this case, the decision line will move such that the overlapped rectangle region belongs to region 1. Hence probability of error is zero. g) Draw the minimax decision surface. Compare and contrast this to your answer in part (c). The requirement for the minimax decision surface is Since P(x/w 1 )=P(x/w 2 ), we need to obtain R 1 = R 2 The minimax decision surface also will have infinite solution compared to Bayes’ decision surface. The contrast is the overlap region needs to be divided into equal area to get R 1 = R 2 LECTURE 15: EXAM NO. 1 (CHAP. 2)