Presentation is loading. Please wait.

Presentation is loading. Please wait.

The UNIVERSITY of Kansas EECS 800 Research Seminar Mining Biological Data Instructor: Luke Huan Fall, 2006.

Similar presentations


Presentation on theme: "The UNIVERSITY of Kansas EECS 800 Research Seminar Mining Biological Data Instructor: Luke Huan Fall, 2006."— Presentation transcript:

1 The UNIVERSITY of Kansas EECS 800 Research Seminar Mining Biological Data Instructor: Luke Huan Fall, 2006

2 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide2 10/23/2006 Classification II Overivew KNN Native Bayes FLD SVM

3 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide3 10/23/2006 Classification II Classification (i.e. Discrimination) CS version: Pattern Recognition Artificial Intelligence Neural Networks Data Mining Machine Learning

4 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide4 10/23/2006 Classification II Classification (i.e. discrimination) Statistics Model Classes with Probability Distribution CS Use to study class difference & find rules Data are just Sets of Numbers Rules distinguish between these Current thought: combine these

5 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide5 10/23/2006 Classification II Classification Basics Point Clouds

6 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide6 10/23/2006 Classification II Instance-Based Classifiers Store the training records Use training records to predict the class label of unseen cases

7 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide7 10/23/2006 Classification II Instance Based Classifiers Examples: Rote-learner Memorizes entire training data and performs classification only if attributes of record match one of the training examples exactly Nearest neighbor Uses k “closest” points (nearest neighbors) for performing classification

8 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide8 10/23/2006 Classification II Nearest Neighbor Classifiers Basic idea: If it walks like a duck, quacks like a duck, then it’s probably a duck Training Records Test Record Compute Distance Choose k of the “nearest” records

9 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide9 10/23/2006 Classification II Nearest-Neighbor Classifiers l Requires three things –The set of stored records –Distance Metric to compute distance between records –The value of k, the number of nearest neighbors to retrieve l To classify an unknown record: –Compute distance to other training records –Identify k nearest neighbors –Use class labels of nearest neighbors to determine the class label of unknown record (e.g., by taking majority vote)

10 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide10 10/23/2006 Classification II Definition of Nearest Neighbor K-nearest neighbors of a record x are data points that have the k smallest distance to x

11 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide11 10/23/2006 Classification II 1 nearest-neighbor Voronoi Diagram

12 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide12 10/23/2006 Classification II Nearest Neighbor Classification Compute distance between two points: Euclidean distance Determine the class from nearest neighbor list take the majority vote of class labels among the k-nearest neighbors Weigh the vote according to distance weight factor, w = 1/d 2

13 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide13 10/23/2006 Classification II Nearest Neighbor Classification… Choosing the value of k: If k is too small, sensitive to noise points If k is too large, neighborhood may include points from other classes

14 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide14 10/23/2006 Classification II Nearest Neighbor Classification… Scaling issues Attributes may have to be scaled to prevent distance measures from being dominated by one of the attributes Example: height of a person may vary from 1.5m to 1.8m weight of a person may vary from 90lb to 300lb income of a person may vary from $10K to $1M

15 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide15 10/23/2006 Classification II Nearest Neighbor Classification… Problem with Euclidean measure: High dimensional data curse of dimensionality Can produce counter-intuitive results 1 1 1 1 1 1 1 1 1 1 1 0 0 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 vs d = 1.4142 Solution: Normalize the vectors to unit length

16 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide16 10/23/2006 Classification II Nearest neighbor Classification… k-NN classifiers are lazy learners It does not build models explicitly Unlike eager learners such as decision tree induction Classifying unknown records are relatively expensive

17 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide17 10/23/2006 Classification II Example: PEBLS PEBLS: Parallel Exemplar-Based Learning System (Cost & Salzberg) Works with both continuous and nominal features For nominal features, distance between two nominal values is computed using modified value difference metric (MVDM) Each record is assigned a weight factor Number of nearest neighbor, k = 1

18 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide18 10/23/2006 Classification II Example: PEBLS Distance between nominal attribute values: d(Single,Married) = | 2/4 – 0/4 | + | 2/4 – 4/4 | = 1 d(Single,Divorced) = | 2/4 – 1/2 | + | 2/4 – 1/2 | = 0 d(Married,Divorced) = | 0/4 – 1/2 | + | 4/4 – 1/2 | = 1 d(Refund=Yes,Refund=No) = | 0/3 – 3/7 | + | 3/3 – 4/7 | = 6/7 Class Marital Status SingleMarriedDivorced Yes201 No241 Class Refund YesNo Yes03 No34

19 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide19 10/23/2006 Classification II Example: PEBLS Distance between record X and record Y: where: w X  1 if X makes accurate prediction most of the time w X > 1 if X is not reliable for making predictions

20 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide20 10/23/2006 Classification II Bayes Classifier A probabilistic framework for solving classification problems Conditional Probability: Bayes theorem:

21 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide21 10/23/2006 Classification II Example of Bayes Theorem Given: A doctor knows that meningitis causes stiff neck 50% of the time Prior probability of any patient having meningitis is 1/50,000 Prior probability of any patient having stiff neck is 1/20 If a patient has stiff neck, what’s the probability he/she has meningitis?

22 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide22 10/23/2006 Classification II Bayesian Classifiers Consider each attribute and class label as random variables Given a record with attributes (A 1, A 2,…,A n ) Goal is to predict class C Specifically, we want to find the value of C that maximizes P(C| A 1, A 2,…,A n ) Can we estimate P(C| A 1, A 2,…,A n ) directly from data?

23 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide23 10/23/2006 Classification II Bayesian Classifiers Approach: compute the posterior probability P(C | A 1, A 2, …, A n ) for all values of C using the Bayes theorem Choose value of C that maximizes P(C | A 1, A 2, …, A n ) Equivalent to choosing value of C that maximizes P(A 1, A 2, …, A n |C) P(C) How to estimate P(A 1, A 2, …, A n | C )?

24 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide24 10/23/2006 Classification II Naïve Bayes Classifier Assume independence among attributes A i when class is given: P(A 1, A 2, …, A n |C) = P(A 1 | C j ) P(A 2 | C j )… P(A n | C j ) Can estimate P(A i | C j ) for all A i and C j. New point is classified to C j if P(C j )  P(A i | C j ) is maximal.

25 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide25 10/23/2006 Classification II How to Estimate Probabilities from Data? Class: P(C) = N c /N e.g., P(No) = 7/10, P(Yes) = 3/10 For discrete attributes: P(A i | C k ) = |A ik |/ N c where |A ik | is number of instances having attribute A i and belongs to class C k Examples: P(Status=Married|No) = 4/7 P(Refund=Yes|Yes)=0 k

26 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide26 10/23/2006 Classification II How to Estimate Probabilities from Data? For continuous attributes: Discretize the range into bins one ordinal attribute per bin violates independence assumption Two-way split: (A v) choose only one of the two splits as new attribute Probability density estimation: Assume attribute follows a normal distribution Use data to estimate parameters of distribution (e.g., mean and standard deviation) Once probability distribution is known, can use it to estimate the conditional probability P(A i |c) k

27 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide27 10/23/2006 Classification II How to Estimate Probabilities from Data? Normal distribution: One for each (A i,c i ) pair For (Income, Class=No): If Class=No sample mean = 110 sample variance = 2975

28 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide28 10/23/2006 Classification II Naïve Bayes Classifier If one of the conditional probability is zero, then the entire expression becomes zero Probability estimation: c: number of classes p: prior probability m: parameter

29 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide29 10/23/2006 Classification II Naïve Bayes (Summary) Robust to isolated noise points Handle missing values by ignoring the instance during probability estimate calculations Robust to irrelevant attributes Independence assumption may not hold for some attributes Use other techniques such as Bayesian Belief Networks (BBN)

30 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide30 10/23/2006 Classification II Revisiting Naïve Bayes Harder Example (slanted clouds):

31 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide31 10/23/2006 Classification II Data Normalization Rescale (standardize) coordinate axes i. e. replace (full) data matrix: Then do mean difference

32 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide32 10/23/2006 Classification II A Toy 2D Case

33 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide33 10/23/2006 Classification II Why Covariance is Important “Orientation” of the distribution is captured by covariance Which one is red? + or

34 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide34 10/23/2006 Classification II Fisher Linear Discrimination Useful notation (data vectors of length ): Class +1: Class -1: Centerpoints: and Based on S. Marron

35 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide35 10/23/2006 Classification II Fisher Linear Discrimination Covariances, for (outer products) Based on centered, normalized data matrices: Note: use “MLE” version of estimated covariance matrices, for simpler notation

36 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide36 10/23/2006 Classification II Assumption Major Assumption: Class covariances are the same (or “similar”) Like this: Not this:

37 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide37 10/23/2006 Classification II Fisher Linear Discrimination Good estimate of (common) within class cov? Pooled (weighted average) within class cov: based on the combined full data matrix:

38 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide38 10/23/2006 Classification II Fisher Linear Discrimination Simple way to find “correct cov. adjustment”: Individually transform subpopulations so “spherical” about their means For define

39 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide39 10/23/2006 Classification II Fisher Linear Discrimination In transformed space, best separating hyperplane is perpendicular bisector of line between means

40 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide40 10/23/2006 Classification II Fisher Linear Discrimination Relationship to Mahalanobis distance Idea: For, a natural distance measure is: “unit free”, i.e. “standardized” essentially mod out covariance structure Euclidean dist. applied to & Same as key transformation for FLD I.e. FLD is mean difference in Mahalanobis space

41 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide41 10/23/2006 Classification II Support Vector Machines Find a linear hyperplane (decision boundary) that will separate the data

42 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide42 10/23/2006 Classification II Support Vector Machines One Possible Solution

43 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide43 10/23/2006 Classification II Support Vector Machines Another possible solution

44 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide44 10/23/2006 Classification II Support Vector Machines Other possible solutions

45 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide45 10/23/2006 Classification II Support Vector Machines Which one is better? B1 or B2? How do you define better?

46 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide46 10/23/2006 Classification II Support Vector Machines Find hyperplane maximizes the margin => B1 is better than B2

47 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide47 10/23/2006 Classification II Support Vector Machines

48 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide48 10/23/2006 Classification II Support Vector Machines We want to maximize: Which is equivalent to minimizing: But subjected to the following constraints: This is a constrained optimization problem –Numerical approaches to solve it (e.g., quadratic programming)

49 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide49 10/23/2006 Classification II Support Vector Machines What if the problem is not linearly separable?

50 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide50 10/23/2006 Classification II Support Vector Machines What if the problem is not linearly separable? Introduce slack variables Need to minimize: Subject to:

51 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide51 10/23/2006 Classification II Nonlinear Support Vector Machines What if decision boundary is not linear?

52 Mining Biological Data KU EECS 800, Luke Huan, Fall’06 slide52 10/23/2006 Classification II Nonlinear Support Vector Machines Transform data into higher dimensional space


Download ppt "The UNIVERSITY of Kansas EECS 800 Research Seminar Mining Biological Data Instructor: Luke Huan Fall, 2006."

Similar presentations


Ads by Google