Discrimination and Classification. Discrimination Situation: We have two or more populations  1,  2, etc (possibly p-variate normal). The populations.

Slides:



Advertisements
Similar presentations
Discrimination amongst k populations. We want to determine if an observation vector comes from one of the k populations For this purpose we need to partition.
Advertisements

SEEM Tutorial 4 – Clustering. 2 What is Cluster Analysis?  Finding groups of objects such that the objects in a group will be similar (or.
Clustering (2). Hierarchical Clustering Produces a set of nested clusters organized as a hierarchical tree Can be visualized as a dendrogram –A tree like.
Hierarchical Clustering
1 CSE 980: Data Mining Lecture 16: Hierarchical Clustering.
Hierarchical Clustering, DBSCAN The EM Algorithm
Data Mining Cluster Analysis: Basic Concepts and Algorithms
Agglomerative Hierarchical Clustering 1. Compute a distance matrix 2. Merge the two closest clusters 3. Update the distance matrix 4. Repeat Step 2 until.
K Means Clustering , Nearest Cluster and Gaussian Mixture
Clustering Clustering of data is a method by which large sets of data is grouped into clusters of smaller sets of similar data. The example below demonstrates.
Pattern Classification, Chapter 2 (Part 2) 0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R.
Pattern Classification. Chapter 2 (Part 1): Bayesian Decision Theory (Sections ) Introduction Bayesian Decision Theory–Continuous Features.
Pattern Classification, Chapter 2 (Part 2) 0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R.
Chapter 2: Bayesian Decision Theory (Part 2) Minimum-Error-Rate Classification Classifiers, Discriminant Functions and Decision Surfaces The Normal Density.
Pattern Classification Chapter 2 (Part 2)0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O.
Topic 6: Introduction to Hypothesis Testing
Classification and risk prediction
CS347 Review Slides (IR Part II) June 6, 2001 ©Prabhakar Raghavan.
Clustering… in General In vector space, clusters are vectors found within  of a cluster vector, with different techniques for determining the cluster.
Today Unsupervised Learning Clustering K-means. EE3J2 Data Mining Lecture 18 K-means and Agglomerative Algorithms Ali Al-Shahib.
0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Ulf Schmitz, Pattern recognition - Clustering1 Bioinformatics Pattern recognition - Clustering Ulf Schmitz
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
1 Linear Classification Problem Two approaches: -Fisher’s Linear Discriminant Analysis -Logistic regression model.
1 Comparison of Discrimination Methods for the Classification of Tumors Using Gene Expression Data Presented by: Tun-Hsiang Yang.
The Neymann-Pearson Lemma Suppose that the data x 1, …, x n has joint density function f(x 1, …, x n ;  ) where  is either  1 or  2. Let g(x 1, …,
Clustering Unsupervised learning Generating “classes”
Clustering Algorithms Mu-Yu Lu. What is Clustering? Clustering can be considered the most important unsupervised learning problem; so, as every other.
Summarized by Soo-Jin Kim
Inference for the mean vector. Univariate Inference Let x 1, x 2, …, x n denote a sample of n from the normal distribution with mean  and variance 
Principles of Pattern Recognition
Chapter 12 – Discriminant Analysis © Galit Shmueli and Peter Bruce 2010 Data Mining for Business Intelligence Shmueli, Patel & Bruce.
Chapter 8 Discriminant Analysis. 8.1 Introduction  Classification is an important issue in multivariate analysis and data mining.  Classification: classifies.
CLUSTERING. Overview Definition of Clustering Existing clustering methods Clustering examples.
CSE5334 DATA MINING CSE4334/5334 Data Mining, Fall 2014 Department of Computer Science and Engineering, University of Texas at Arlington Chengkai Li (Slides.
Linear Discriminant Analysis and Its Variations Abu Minhajuddin CSE 8331 Department of Statistical Science Southern Methodist University April 27, 2002.
Clustering.
Clustering Algorithms Presented by Michael Smaili CS 157B Spring
Chapter 3: Maximum-Likelihood Parameter Estimation l Introduction l Maximum-Likelihood Estimation l Multivariate Case: unknown , known  l Univariate.
K-Means Algorithm Each cluster is represented by the mean value of the objects in the cluster Input: set of objects (n), no of clusters (k) Output:
CS 8751 ML & KDDData Clustering1 Clustering Unsupervised learning Generating “classes” Distance/similarity measures Agglomerative methods Divisive methods.
Hierarchical Clustering Produces a set of nested clusters organized as a hierarchical tree Can be visualized as a dendrogram – A tree like diagram that.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Discrimination and Classification. Discrimination Situation: We have two or more populations  1,  2, etc (possibly p-variate normal). The populations.
Inference for the mean vector. Univariate Inference Let x 1, x 2, …, x n denote a sample of n from the normal distribution with mean  and variance 
Classification Course web page: vision.cis.udel.edu/~cv May 14, 2003  Lecture 34.
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Linear Classifier Team teaching.
CLUSTER ANALYSIS. Cluster Analysis  Cluster analysis is a major technique for classifying a ‘mountain’ of information into manageable meaningful piles.
Lecture 2. Bayesian Decision Theory
Chapter 12 – Discriminant Analysis
Data Mining: Basic Cluster Analysis
Chapter 3: Maximum-Likelihood Parameter Estimation
Inference for the mean vector
Discrimination and Classification
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
A graphical explanation
Mathematical Foundations of BME
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Multivariate Methods Berlin Chen
BIRCH: Balanced Iterative Reducing and Clustering Using Hierarchies
Multivariate Methods Berlin Chen, 2005 References:
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
The two sample problem.
SEEM4630 Tutorial 3 – Clustering.
EM Algorithm and its Applications
Discrimination and Classification
Presentation transcript:

Discrimination and Classification

Discrimination Situation: We have two or more populations  1,  2, etc (possibly p-variate normal). The populations are known (or we have data from each population) We have data for a new case (population unknown) and we want to identify the which population for which the new case is a member.

The Basic Problem Suppose that the data from a new case x 1, …, x p has joint density function either :  1 : g(x 1, …, x n ) or  2 : h(x 1, …, x n ) We want to make the decision to D 1 : Classify the case in  1 (g is the correct distribution) or D 2 : Classify the case in  2 (h is the correct distribution)

The Two Types of Errors 1.Misclassifying the case in  1 when it actually lies in  2. Let P[1|2] = P[D 1 |  2 ] = probability of this type of error 2.Misclassifying the case in  2 when it actually lies in  1. Let P[2|1] = P[D 2 |  1 ] = probability of this type of error This is similar Type I and Type II errors in hypothesis testing.

Note: 1. C 1 = the region were we make the decision D 1. (the decision to classify the case in  1 ) A discrimination scheme is defined by splitting p – dimensional space into two regions. 2. C 2 = the region were we make the decision D 2. (the decision to classify the case in  2 )

1.Set up the regions C 1 and C 2 so that one of the probabilities of misclassification, P[2|1] say, is at some low acceptable value . Accept the level of the other probability of misclassification P[1|2] = . There can be several approaches to determining the regions C 1 and C 2. All concerned with taking into account the probabilities of misclassification P[2|1] and P[1|2]

2.Set up the regions C 1 and C 2 so that the total probability of misclassification: P[Misclassification] = P[1] P[2|1] + P[2]P[1|2] is minimized P[1] = P[the case belongs to  1 ] P[2] = P[the case belongs to  2 ]

3.Set up the regions C 1 and C 2 so that the total expected cost of misclassification: E[Cost of Misclassification] = ECM = c 2|1 P[1] P[2|1] + c 1|2 P[2]P[1|2] is minimized P[1] = P[the case belongs to  1 ] P[2] = P[the case belongs to  2 ] c 2|1 = the cost of misclassifying the case in  2 when the case belongs to  1. c 1|2 = the cost of misclassifying the case in  1 when the case belongs to  2.

The Optimal Classification Rule Suppose that the data x 1, …, x p has joint density function f(x 1, …, x p ;  ) where  is either  1 or  2. Let g(x 1, …, x p ) = f(x 1, …, x n ;  1 ) and h(x 1, …, x p ) = f(x 1, …, x n ;  2 ) We want to make the decision D 1 :  =  1 (g is the correct distribution) against D 2 :  =  2 (h is the correct distribution)

and where then the optimal regions (minimizing ECM, expected cost of misclassification) for making the decisions D 1 and D 2 respectively are C 1 and C 2

Proof: ECM = E[Cost of Misclassification] = c 2|1 P[1] P[2|1] + c 1|2 P[2]P[1|2]

Therefore Thus ECM is minimized if C 1 contains all of the points (x 1, …, x p ) such that the integrand is negative

Fishers Linear Discriminant Function. Suppose that x 1, …, x p is either data from a p-variate Normal distribution with mean vector: The covariance matrix  is the same for both populations  1 and  2.

The Neymann-Pearson Lemma states that we should classify into populations  1 and  2 using: That is make the decision D 1 : population is  1 if > k

or and

Finally we make the decision D 1 : population is  1 if where and Note: k = 1 and ln k = 0 if c 1|2 = c 2|1 and P[1] = P[2].

The function Is called Fisher’s linear discriminant function

In the case where the populations are unknown but estimated from data Fisher’s linear discriminant function

Example 2 Annual financial data are collected for firms approximately 2 years prior to bankruptcy and for financially sound firms at about the same point in time. The data on the four variables x 1 = CF/TD = (cash flow)/(total debt), x 2 = NI/TA = (net income)/(Total assets), x 3 = CA/CL = (current assets)/(current liabilties, and x 4 = CA/NS = (current assets)/(net sales) are given in the following table.

The data are given in the following table:

Examples using SPSS

Classification or Cluster Analysis Have data from one or several populations

Situation Have multivariate (or univariate) data from one or several populations (the number of populations is unknown) Want to determine the number of populations and identify the populations

Example

Hierarchical Clustering Methods The following are the steps in the agglomerative Hierarchical clustering algorithm for grouping N objects (items or variables). 1.Start with N clusters, each consisting of a single entity and an N X N symmetric matrix (table) of distances (or similarities) D = (d ij ). 2.Search the distance matrix for the nearest (most similar) pair of clusters. Let the distance between the "most similar" clusters U and V be d UV. 3.Merge clusters U and V. Label the newly formed cluster (UV). Update the entries in the distance matrix by a)deleting the rows and columns corresponding to clusters U and V and b)adding a row and column giving the distances between cluster (UV) and the remaining clusters.

4.Repeat steps 2 and 3 a total of N-1 times. (All objects will be a single cluster a termination of this algorithm.) Record the identity of clusters that are merged and the levels (distances or similarities) at which the mergers take place.

Different methods of computing inter-cluster distance

Example To illustrate the single linkage algorithm, we consider the hypothetical distance matrix between pairs of five objects given below:

Treating each object as a cluster, the clustering begins by merging the two closest items (3 & 5). To implement the next level of clustering we need to compute the distances between cluster (35) and the remaining objects: d (35)1 = min{3,11} = 3 d (35)2 = min{7,10} = 7 d (35)4 = min{9,8} = 8 The new distance matrix becomes:

The next two closest clusters ((35) & 1) are merged to form cluster (135). Distances between this cluster and the remaining clusters become:

Distances between this cluster and the remaining clusters become: d (135)2 = min{7,9} = 7 d (135)4 = min{8,6} = 6 The distance matrix now becomes: Continuing the next two closest clusters (2 & 4) are merged to form cluster (24).

Distances between this cluster and the remaining clusters become: d (135)(24) = min{d (135)2,d (135)4 )= min{7,6} = 6 The final distance matrix now becomes: At the final step clusters (135) and (24) are merged to form the single cluster (12345) of all five items.

The results of this algorithm can be summarized graphically on the following "dendogram"

Dendograms for clustering the 11 languages on the basis of the ten numerals

Dendogram Cluster Analysis of N=22 Utility companies Euclidean distance, Average Linkage

Dendogram Cluster Analysis of N=22 Utility companies Euclidean distance, Single Linkage