COMP 328: Final Review Spring 2010 Nevin L. Zhang Department of Computer Science & Engineering The Hong Kong University of Science & Technology

Slides:



Advertisements
Similar presentations
University of Joensuu Dept. of Computer Science P.O. Box 111 FIN Joensuu Tel fax Gaussian Mixture.
Advertisements

K Means Clustering , Nearest Cluster and Gaussian Mixture
An Overview of Machine Learning
Supervised Learning Recap
Model-based clustering of gene expression data Ka Yee Yeung 1,Chris Fraley 2, Alejandro Murua 3, Adrian E. Raftery 2, and Walter L. Ruzzo 1 1 Department.
Visual Recognition Tutorial
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
EE 290A: Generalized Principal Component Analysis Lecture 6: Iterative Methods for Mixture-Model Segmentation Sastry & Yang © Spring, 2011EE 290A, University.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
COMP 328: Midterm Review Spring 2010 Nevin L. Zhang Department of Computer Science & Engineering The Hong Kong University of Science & Technology
Announcements  Project proposal is due on 03/11  Three seminars this Friday (EB 3105) Dealing with Indefinite Representations in Pattern Recognition.
Cluster Analysis.  What is Cluster Analysis?  Types of Data in Cluster Analysis  A Categorization of Major Clustering Methods  Partitioning Methods.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Clustering.
Expectation Maximization for GMM Comp344 Tutorial Kai Zhang.
Visual Recognition Tutorial
Revision (Part II) Ke Chen COMP24111 Machine Learning Revision slides are going to summarise all you have learnt from Part II, which should be helpful.
Clustering with Bregman Divergences Arindam Banerjee, Srujana Merugu, Inderjit S. Dhillon, Joydeep Ghosh Presented by Rohit Gupta CSci 8980: Machine Learning.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Midterm Review. 1-Intro Data Mining vs. Statistics –Predictive v. experimental; hypotheses vs data-driven Different types of data Data Mining pitfalls.
Predicting Income from Census Data using Multiple Classifiers Presented By: Arghya Kusum Das Arnab Ganguly Manohar Karki Saikat Basu Subhajit Sidhanta.
Data mining and machine learning A brief introduction.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
CSC321: Neural Networks Lecture 12: Clustering Geoffrey Hinton.
1 HMM - Part 2 Review of the last lecture The EM algorithm Continuous density HMM.
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 16 Nov, 3, 2011 Slide credit: C. Conati, S.
CHAPTER 7: Clustering Eick: K-Means and EM (modified Alpaydin transparencies and new transparencies added) Last updated: February 25, 2014.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Ch 4. Linear Models for Classification (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized and revised by Hee-Woong Lim.
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
Overview of the final test for CSC Overview PART A: 7 easy questions –You should answer 5 of them. If you answer more we will select 5 at random.
Clustering Algorithms Presented by Michael Smaili CS 157B Spring
HMM - Part 2 The EM algorithm Continuous density HMM.
Cluster Analysis Potyó László. Cluster: a collection of data objects Similar to one another within the same cluster Similar to one another within the.
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
Lecture 2: Statistical learning primer for biologists
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
CSE 446 Dimensionality Reduction and PCA Winter 2012 Slides adapted from Carlos Guestrin & Luke Zettlemoyer.
Flat clustering approaches
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Classification Ensemble Methods 1
Model-based Clustering
Information Retrieval Search Engine Technology (8) Prof. Dragomir R. Radev.
1 Cluster Analysis – 2 Approaches K-Means (traditional) Latent Class Analysis (new) by Jay Magidson, Statistical Innovations based in part on a presentation.
Bayesian Hierarchical Clustering Paper by K. Heller and Z. Ghahramani ICML 2005 Presented by David Williams Paper Discussion Group ( )
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Gaussian Mixture Model classification of Multi-Color Fluorescence In Situ Hybridization (M-FISH) Images Amin Fazel 2006 Department of Computer Science.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
CMPS 142/242 Review Section Fall 2011 Adapted from Lecture Slides.
Big Data Infrastructure Week 9: Data Mining (4/4) This work is licensed under a Creative Commons Attribution-Noncommercial-Share Alike 3.0 United States.
Unsupervised Learning Part 2. Topics How to determine the K in K-means? Hierarchical clustering Soft clustering with Gaussian mixture models Expectation-Maximization.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
A Brief Introduction to Bayesian networks
Clustering (1) Clustering Similarity measure Hierarchical clustering
LECTURE 11: Advanced Discriminant Analysis
Classification of unlabeled data:
The Elements of Statistical Learning
Clustering Evaluation The EM Algorithm
Latent Variables, Mixture Models and EM
SEG 4630 E-Commerce Data Mining — Final Review —
Probabilistic Models with Latent Variables
Revision (Part II) Ke Chen
Revision (Part II) Ke Chen
Unsupervised Learning II: Soft Clustering with Gaussian Mixture Models
Hairong Qi, Gonzalez Family Professor
What is Artificial Intelligence?
Presentation transcript:

COMP 328: Final Review Spring 2010 Nevin L. Zhang Department of Computer Science & Engineering The Hong Kong University of Science & Technology Can be used as cheat sheet

Page 2 Pre-Midterm l Algorithms for supervised learning n Decision trees n Instance-based learning n Naïve Bayes classifiers n Neural networks n Support vector machines l General issues regarding supervised learning n Classification error and confidence interval n Bias-Variance tradeoff n PAC learning theory

Post-Midterm l Clustering n Distance-Based Clustering n Model-Based Clustering l Dimension Reduction n Principal Component Analysis l Reinforcement Learning l Ensemble Learning

Clustering

Distance/Similarity Measures

Distance-Based Clustering l Partitional and Hierarchical clustering

K-Means: Partitional Clustering

l Different initial points might lead to different partitions l Solution: n Multiple runs, n Use evaluation criteria such as SSE to pick the best one

Hierarchical Clustering l Agglomerative and Divisive

Cluster Similarity

Cluster Validation l External indices n Entropy: Average purity of clusters obtained n Mutual Information between class label and cluster label

Cluster Validation l External Measure n Jaccard Index n Rand Index Measure similarity between two relationships: in-same-class & in-same-cluster # pairs in same cluster# pairs in diff cluster # pairs w/ same labelab # pairs w/ diff labelcd

Cluster Validation l Internal Measure n Dunn’s index

Cluster Validation l Internal Measure

Post-Midterm l Clustering n Distance-Based Clustering n Model-Based Clustering l Dimension Reduction n Principal Component Analysis l Reinforcement Learning l Ensemble Learning

Model-Based Clustering l Assume data generated from a mixture model with K components l Estimate parameters of the model from data l Assign objects to clusters based posterior probability: Soft Assignment

Gaussian Mixtures

Learning Gaussian Mixture Models

EM

l l(t): Log likelihood of model after t-th iteration l l(t): increases monotonically with t l But might go to infinite in case of singularity n Solution: place bound on eigen values of covariance matrix l Local maximum n Multiple restart n Use likelihood to pick best model

EM and K-Means l K-Means is hard-assignment EM

Mixture Variable for Discrete Data

Latent Class Model

Learning Latent Class Models Always converges

Post-Midterm l Clustering n Distance-Based Clustering n Model-Based Clustering l Dimension Reduction n Principal Component Analysis l Reinforcement Learning l Ensemble Learning

Dimension Reduction l Necessary because there are data sets with large numbers of attributes that are difficult to learning algorithms to handle.

Principal Component Analysis

PCA Solution

PCA Illustration

Eigenvalues and Projection Error

Post-Midterm l Clustering n Distance-Based Clustering n Model-Based Clustering l Dimension Reduction n Principal Component Analysis l Reinforcement Learning l Ensemble Learning

Reinforcement Learning

Markov Decision Process l A model of how agent interact with its environment

Markov Decision Process

Value Iteration

Reinforcement Learning

Q-Learning

l From Q-function based value iteration l Ideas n In-place/asynchronous value iteration n Approximate expectation using samples n ε-greedy policy (for exploration/exploitation) tradeoff

Time Difference Learning

Sarsa is also time difference learning

Post-Midterm l Clustering n Distance-Based Clustering n Model-Based Clustering l Dimension Reduction n Principal Component Analysis l Reinforcement Learning l Ensemble Learning

Ensemble Learning

Bagging: Reduce Variance

Boosting: Reduce Classification Error

AdaBoost: Exponential Error