Project 2 k-NN 2018/11/10.

Slides:



Advertisements
Similar presentations
DECISION TREES. Decision trees  One possible representation for hypotheses.
Advertisements

Face Recognition and Biometric Systems Eigenfaces (2)
Notes Sample vs distribution “m” vs “µ” and “s” vs “σ” Bias/Variance Bias: Measures how much the learnt model is wrong disregarding noise Variance: Measures.
$100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300.
K nearest neighbor and Rocchio algorithm
Mapping Between Taxonomies Elena Eneva 30 Oct 2001 Advanced IR Seminar.
CS Team 5 Alex Wong Raheel Khan Rumeiz Hasseem Swati Bharati Biometric Authentication System.
Learning from Observations Chapter 18 Section 1 – 4.
More Methodology; Nearest-Neighbor Classifiers Sec 4.7.
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Carla P. Gomes Module: Nearest Neighbor Models (Reading: Chapter.
Classification Dr Eamonn Keogh Computer Science & Engineering Department University of California - Riverside Riverside,CA Who.
Chapter 2: Pattern Recognition
Proteomic Mass Spectrometry
Bregman Divergences in Clustering and Dimensionality Reduction COMS : Learning and Empirical Inference Irina Rish IBM T.J. Watson Research Center.
These slides are based on Tom Mitchell’s book “Machine Learning” Lazy learning vs. eager learning Processing is delayed until a new instance must be classified.
1 Nearest Neighbor Learning Greg Grudic (Notes borrowed from Thomas G. Dietterich and Tom Mitchell) Intro AI.
Steep learning curves Reading: DH&S, Ch 4.6, 4.5.
Making the Most of Small Sample High Dimensional Micro-Array Data Allan Tucker, Veronica Vinciotti, Xiaohui Liu; Brunel University Paul Kellam; Windeyer.
Nearest Neighbor Classifiers other names: –instance-based learning –case-based learning (CBL) –non-parametric learning –model-free learning.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
$100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Digital Image Processing & Pattern Analysis (CSCE 563) Introduction to Pattern Analysis Prof. Amr Goneid Department of Computer Science & Engineering The.
Weka Just do it Free and Open Source ML Suite Ian Witten & Eibe Frank University of Waikato New Zealand.
ECE 471/571 - Lecture 19 Review 11/12/15. A Roadmap 2 Pattern Classification Statistical ApproachNon-Statistical Approach SupervisedUnsupervised Basic.
CS Machine Learning Instance Based Learning (Adapted from various sources)
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Instance-Based Learning Evgueni Smirnov. Overview Instance-Based Learning Comparison of Eager and Instance-Based Learning Instance Distances for Instance-Based.
Презентацию подготовила Хайруллина Ч.А. Муслюмовская гимназия Подготовка к части С ЕГЭ.
Defect Prediction Techniques He Qing What is Defect Prediction? Use historical data to predict defect. To plan for allocation of defect detection.
PLUS.
CS : Designing, Visualizing and Understanding Deep Neural Networks
Septian Adi Wijaya – Informatics Brawijaya University
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
ECE 471/571 - Lecture 19 Review 02/24/17.
Performance Evaluation 02/15/17
Classification Nearest Neighbor
Chapter 7 – K-Nearest-Neighbor
Project 4 User and Movie 2018/11/10.
Instance Based Learning (Adapted from various sources)
Nearest-Neighbor Classifiers
Project 1 Binary Classification
Text Categorization Rong Jin.
+ + Your text Enter your text Please enter your text Your text
ECE 471/571 – Review 1.
Parametric Estimation
COSC 4335: Other Classification Techniques
Junheng, Shengming, Yunsheng 11/2/2018
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Intro to Machine Learning
Extracting Recipes from Chemical Academic Papers
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Test Process “V” Diagram
Information Retrieval
CSE4334/5334 Data Mining Lecture 7: Classification (4)
Project 3 Image Retrieval
CLICK TO ADD TITLE IN HERE
Introduction.
Standards of Measurement
[replace this text] [replace this text]
ECE – Pattern Recognition Lecture 10 – Nonparametric Density Estimation – k-nearest-neighbor (kNN) Hairong Qi, Gonzalez Family Professor Electrical.
4 Step Simple Diagram Sample text Sample text Sample text Sample text
PLEASE ADD YOUR TITLE HERE.
Bayesian Decision Theory
Hairong Qi, Gonzalez Family Professor
ECE – Pattern Recognition Lecture 8 – Performance Evaluation
ECE – Pattern Recognition Lecture 4 – Parametric Estimation
单击此处添加文字标题 单击此处添加副标题.
Hairong Qi, Gonzalez Family Professor
ECE – Pattern Recognition Midterm Review
Presentation transcript:

Project 2 k-NN 2018/11/10

Data Data: wine.data (can be opened with Text Editors, e.g. Notepad, Word, …) These data came from the chemical analysis results of different categories of wine For each sample, 13 attributes are provided Please read the readme.txt carefully Source: http://archive.ics.uci.edu/ml/datasets/Wine Problem: classify the wine into one of three categories 2018/11/10

Requirement Basic (7/10) Plus (at most 3/10) 1) Develop a k-NN classifier with Euclidean distance and simple voting 2) Perform 5-fold cross validation, find out which k performs the best (in terms of accuracy) 3) Use PCA to reduce the dimensionality, then perform 2) again. Does PCA improve the accuracy? Which dimensionality is better? Plus (at most 3/10) Try other distance metrics or distance-based voting (+0.5~1) Try other dimensionality reduction methods (+1~2) Perform feature selection (+1) How to set the k value, if not using cross validation? Verify your idea (+3) 2018/11/10

Report .pdf is best, .doc and .ppt are acceptable English is encouraged List all references Source code is a plus 2018/11/10