ECE 471/571 - Lecture 19 Review 11/12/15. A Roadmap 2 Pattern Classification Statistical ApproachNon-Statistical Approach SupervisedUnsupervised Basic.

Slides:



Advertisements
Similar presentations
ECE 471/571 - Lecture 13 Gradient Descent 10/13/14.
Advertisements

Unsupervised Learning Clustering K-Means. Recall: Key Components of Intelligent Agents Representation Language: Graph, Bayes Nets, Linear functions Inference.
An Overview of Machine Learning
Supervised Learning Recap
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
Classification and risk prediction
L15:Microarray analysis (Classification). The Biological Problem Two conditions that need to be differentiated, (Have different treatments). EX: ALL (Acute.
Math 5364 Notes Chapter 5: Alternative Classification Techniques Jesse Crawford Department of Mathematics Tarleton State University.
Revision (Part II) Ke Chen COMP24111 Machine Learning Revision slides are going to summarise all you have learnt from Part II, which should be helpful.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Midterm Review. 1-Intro Data Mining vs. Statistics –Predictive v. experimental; hypotheses vs data-driven Different types of data Data Mining pitfalls.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
嵌入式視覺 Pattern Recognition for Embedded Vision Template matching Statistical / Structural Pattern Recognition Neural networks.
This week: overview on pattern recognition (related to machine learning)
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
Machine Learning Lecture 11 Summary G53MLE | Machine Learning | Dr Guoping Qiu1.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Ch 4. Linear Models for Classification (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized and revised by Hee-Woong Lim.
ECE 471/571 – Lecture 2 Bayesian Decision Theory 08/25/15.
ECE 471/571 – Lecture 6 Dimensionality Reduction – Fisher’s Linear Discriminant 09/08/15.
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN © The MIT Press, Lecture.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
KNN & Naïve Bayes Hongning Wang Today’s lecture Instance-based classifiers – k nearest neighbors – Non-parametric learning algorithm Model-based.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall DM Finals Study Guide Rodney Nielsen.
Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.
Elements of Pattern Recognition CNS/EE Lecture 5 M. Weber P. Perona.
ECE 471/571 – Lecture 13 Use Neural Networks for Pattern Recognition 10/13/15.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 9: Review.
ECE 471/571 – Lecture 21 Syntactic Pattern Recognition 11/19/15.
ECE 471/571 – Lecture 3 Discriminant Function and Normal Density 08/27/15.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Supervise Learning. 2 What is learning? “Learning denotes changes in a system that... enable a system to do the same task more efficiently the next time.”
KNN & Naïve Bayes Hongning Wang
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
High resolution product by SVM. L’Aquila experience and prospects for the validation site R. Anniballe DIET- Sapienza University of Rome.
Use Neural Networks for Pattern Recognition 03/29/17
Syntactic Pattern Recognition 04/28/17
ECE 471/571 - Lecture 19 Review 02/24/17.
Nonparametric Density Estimation – k-nearest neighbor (kNN) 02/20/17
Performance Evaluation 02/15/17
ECE 471/571 – Lecture 18 Classifier Fusion 04/12/17.
Unsupervised Learning - Clustering 04/03/17
Unsupervised Learning - Clustering
Pattern Recognition CS479/679 Pattern Recognition Dr. George Bebis
Course Outline MODEL INFORMATION COMPLETE INCOMPLETE
ECE 471/571 – Lecture 12 Perceptron.
ECE 471/571 – Review 1.
Parametric Estimation
Hairong Qi, Gonzalez Family Professor
INTRODUCTION TO Machine Learning
ECE – Pattern Recognition Lecture 16: NN – Back Propagation
ECE – Pattern Recognition Lecture 15: NN - Perceptron
ECE – Pattern Recognition Lecture 13 – Decision Tree
Hairong Qi, Gonzalez Family Professor
Machine Learning Perceptron: Linearly Separable Supervised Learning
Assignment 1: Classification by K Nearest Neighbors (KNN) technique
Hairong Qi, Gonzalez Family Professor
CAMCOS Report Day December 9th, 2015 San Jose State University
ECE – Lecture 1 Introduction.
ECE – Pattern Recognition Lecture 10 – Nonparametric Density Estimation – k-nearest-neighbor (kNN) Hairong Qi, Gonzalez Family Professor Electrical.
Bayesian Decision Theory
Hairong Qi, Gonzalez Family Professor
ECE – Pattern Recognition Lecture 8 – Performance Evaluation
ECE – Pattern Recognition Lecture 4 – Parametric Estimation
Hairong Qi, Gonzalez Family Professor
ECE – Pattern Recognition Lecture 14 – Gradient Descent
ECE – Pattern Recognition Midterm Review
Presentation transcript:

ECE 471/571 - Lecture 19 Review 11/12/15

A Roadmap 2 Pattern Classification Statistical ApproachNon-Statistical Approach SupervisedUnsupervised Basic concepts: Distance Agglomerative method Basic concepts: Baysian decision rule (MPP, LR, Discri.) Parameter estimate (ML, BL) Non-Parametric learning (kNN) LDF (Perceptron) k-means Winner-take-all Kohonen maps Dimensionality Reduction FLD, PCA Performance Evaluation ROC curve ( TP, TN, FN, FP ) cross validation Classifier Fusion majority voting NB, BKS Stochastic Methods local opt (GD) global opt (SA, GA) Decision-tree Syntactic approach NN (BP, Hopfield, DL) Support Vector Machine

3 Q&A What is the common feature of statistical pattern classification? What is supervised pattern classification? How is it different from unsupervised classification? When performing classification, what is the general rule that Baysian decision methods follow? (MPP, Likelihood ratio, Discriminant function) What is the difference between parametric classification and non-parametric classification? What is maximum likelihood estimation? Can you verbally describe kNN? And possible improvements? Both FLD and PCA reduce the dimension of a dataset. However, different projection directions are generated from the two approaches. Comment on the fundamental differences. Learn to plot and use and explain an ROC curve Be familiar with the relationship between TP, FP, TN, FN, precision, recall, specificity, sensitivity, and accuracy

4 Q&A - Cont ’ d What ’ s the difference between feed-forward and feed-backward NN? Examples of each? What is LDF? How to extend from 2-class case to multi-class separation? What ’ s the problem with c 2-class separation and c(c-1)/2 2-class separation? Performance wise, what is the difference between perceptron and bp? Pros and cons How do they work? (need to know the detailed procedure) (no proof needed) Know of different ways to improve the performance of bp  E.g. what is sigmoid? Why is it better?  What is optimal learning rate? Why? The biological structure of a neuron

Q&A – Cont’d Stochastic method What is gradient descent? Know the detailed algorithm. What ’ s the problem? How do SA and GA achieve global optimum as compared to the principle of GD? GA (natural selection with binary tournament mating, exploration with recombination and mutation) Clustering What is unsupervised learning? What i s the difference between dmin, dmax, and dmean? What is agglomerative clustering? Understand k-means, winner-take-all, and Kohonen maps Performance evaluation approaches What is confusion matrix? Fusion Majority voting Naïve Bayes 5

6 Forming Objective Functions Maximum a-posteriori probability Maximum likelihood estimate Fisher ’ s linear discriminant Principal component analysis kNN Perceptron Back propagation neural network Hopfield network ROC curve Various clustering algorithm NB

7 Distance Metrics Minkowski distance Manhattan distance (city-block distance) Euclidean distance Mahalanobis distance Distance between clusters (dmin, dmax, vs. dmean) Classify the following samples into two clusters using dmin, dmax, and dmean. The samples are (0,1), (0,2), (0,3), (0,4), (0,5), (0,6), (0,7), (0,8)

Course evaluation (5 pts) counted in Test 2. 8