Element 2: Discuss basic computational intelligence methods.

Slides:



Advertisements
Similar presentations
Component Analysis (Review)
Advertisements

Data Mining Feature Selection. Data reduction: Obtain a reduced representation of the data set that is much smaller in volume but yet produces the same.
Automatic classification of weld cracks using artificial intelligence and statistical methods Ryszard SIKORA, Piotr BANIUKIEWICZ, Marcin CARYK Szczecin.
Machine Learning Lecture 8 Data Processing and Representation
Data preprocessing before classification In Kennedy et al.: “Solving data mining problems”
Information Retrieval in Practice
Face Recognition Under Varying Illumination Erald VUÇINI Vienna University of Technology Muhittin GÖKMEN Istanbul Technical University Eduard GRÖLLER Vienna.
© 2003 by Davi GeigerComputer Vision September 2003 L1.1 Face Recognition Recognized Person Face Recognition.
Principal Component Analysis
Student Mini-Camp Project Report Pattern Recognition Participant StudentsAffiliations Patrick ChoiClaremont Graduate University Joseph McGrathUniv. of.
Pattern Recognition Topic 1: Principle Component Analysis Shapiro chap
Statistical Methods Chichang Jou Tamkang University.
A quick introduction to the analysis of questionnaire data John Richardson.
Statistical Methods in Computer Science Hypothesis Testing I: Treatment experiment designs Ido Dagan.
CSE 300: Software Reliability Engineering Topics covered: Software metrics and software reliability Software complexity and software quality.
Oral Defense by Sunny Tang 15 Aug 2003
Overview of Search Engines
Discriminant Analysis Testing latent variables as predictors of groups.
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
Jeopardy Hypothesis Testing T-test Basics T for Indep. Samples Z-scores Probability $100 $200$200 $300 $500 $400 $300 $400 $300 $400 $500 $400.
Summarized by Soo-Jin Kim
Enhancing Tensor Subspace Learning by Element Rearrangement
: Chapter 1: Introduction 1 Montri Karnjanadecha ac.th/~montri Principles of Pattern Recognition.
This week: overview on pattern recognition (related to machine learning)
A P STATISTICS LESSON 3 – 2 CORRELATION.
Model of Prediction Error in Chaotic and Web Driven Business Environment Franjo Jović*, Alan Jović ** * Faculty of Electrical Engineering, University of.
Presented by Tienwei Tsai July, 2005
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
BACKGROUND LEARNING AND LETTER DETECTION USING TEXTURE WITH PRINCIPAL COMPONENT ANALYSIS (PCA) CIS 601 PROJECT SUMIT BASU FALL 2004.
General Tensor Discriminant Analysis and Gabor Features for Gait Recognition by D. Tao, X. Li, and J. Maybank, TPAMI 2007 Presented by Iulian Pruteanu.
Access Control Via Face Recognition Progress Review.
Data Mining Chapter 1 Introduction -- Basic Data Mining Tasks -- Related Concepts -- Data Mining Techniques.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Example 1: page 161 #5 Example 2: page 160 #1 Explanatory Variable - Response Variable - independent variable dependent variable.
SINGULAR VALUE DECOMPOSITION (SVD)
 Muhamad Jantan & T. Ramayah School of Management, Universiti Sains Malaysia Data Analysis Using SPSS.
1 Opinion Retrieval from Blogs Wei Zhang, Clement Yu, and Weiyi Meng (2007 CIKM)
Machine Learning Extract from various presentations: University of Nebraska, Scott, Freund, Domingo, Hong,
Chapter 9: Correlation and Regression Analysis. Correlation Correlation is a numerical way to measure the strength and direction of a linear association.
Correlation. Correlation is a measure of the strength of the relation between two or more variables. Any correlation coefficient has two parts – Valence:
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
MACHINE LEARNING 7. Dimensionality Reduction. Dimensionality of input Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Principal Component Analysis (PCA)
CORRELATION ANALYSIS.
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
2D-LDA: A statistical linear discriminant analysis for image matrix
Tallahassee, Florida, 2016 CIS4930 Introduction to Data Mining Midterm Review Peixiang Zhao.
Unsupervised Learning II Feature Extraction
Data statistics and transformation revision Michael J. Watts
Information Retrieval in Practice
Principal Component Analysis (PCA)
Dimensionality Reduction
School of Computer Science & Engineering
Traffic Sign Recognition Using Discriminative Local Features Andrzej Ruta, Yongmin Li, Xiaohui Liu School of Information Systems, Computing and Mathematics.
Principal Component Analysis (PCA)
Machine Learning Ali Ghodsi Department of Statistics
Final Year Project Presentation --- Magic Paint Face
Principal Component Analysis
Outline Peter N. Belhumeur, Joao P. Hespanha, and David J. Kriegman, “Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection,”
PCA vs ICA vs LDA.
Mean transform , a tutorial
CORRELATION ANALYSIS.
Introduction PCA (Principal Component Analysis) Characteristics:
Dimensionality Reduction
Feature Selection Methods
Yamanishi, M., Itoh, M., Kanehisa, M.
Research Institute for Future Media Computing
Principal Component Analysis (PCA)
Unsupervised Learning
Presentation transcript:

Element 2: Discuss basic computational intelligence methods

Feature Extraction Methods 1.Feature Extraction Introduction 2.T-test 3.Signal Noise Ratio 4.Linear correlation coefficient 5.Principle Component Analysis (PCA) 6.Linear Discriminant Analysis (LDA)

Feature Extraction: Definition  When the input data to a system is too large to be processed and it is suspected to be notoriously redundant (much data, but not much information)  The input data will be transformed into a reduced representation set of data.  The transforming is called feature extraction. The obtained reduced set of data is called feature, (also named features vector).

Feature Extraction: Motivation Data compression: Efficient storage Data characterization –Data understanding: analysis Discovering data characteristics –Clustering: unknown labels –Classification: known labels –Pre-processing for further analysis Tracking Visualization: reduction of visual clutter Comparison/classification Search: large collections of data sets Database management: efficient retrieval 4

Feature Extraction Applications Activity recognition Place tracking Face recognition Remote sensing Bioinformatics Structural engineering Robotics Biometrics GIS (Geographic information system) Semiconductor defect analysis Earthquake engineering Plant biology Medicine Sensing … 5

t-Test t-Test is a calculated ranking number for each variable to define how well this variable discriminates two classes. Given two class samples on selected variable Compute t as Class 1 (n 1 samples) Class 2 (n 2 samples)

t-Test Where and are the mean values for this variable for the samples from class 1 and class 2 respectively And s 1 and s 2 are the corresponding sample variance.

Spelling Test Scores A t-test allows us to compare the means of two groups and determine how likely the difference between the two means occurred by chance. The calculations for a t-test requires three pieces of information: - the difference between the means (mean difference) s 1 and s 2 the variance for each group n 1 and n 2 the number of subjects in each group T-test

t-test Example Class 1Class

On Class Practice Try to calculate t value using Excel We have 49 X 1 against 47 X 2 Data File: Moodle – ISCG8042 – Topic 2 - TtestSample.xlsx (worksheet2 ‘Try do this’) Example: worksheet1 ‘example’

On Class Practice con. What’s your result?

Signal Noise Ratio SNR is a calculated ranking number for each variable to define how well this variable discriminates two classes. The following formula is used: where:  1 and  2 are the corresponding standard deviations.

Signal Noise Ratio Iris SNR by NecComBreast-w SNR by NecCom

Linear correlation coefficient (LCC) LCC is a measurement of the strength of a linear relationship between a dependent variable (i.e. the output class, y) and an independent variable (i.e. feature, x) The correlation value varies from –1 to 1. A value of 0 suggests no linear correlation, while values nearer to –1 or 1 mean negatively or positively correlated variables.

Linear correlation coefficient (LCC)

On Class Practice Try to calculate SNR and LCC Data File: Moodle – ISCG8042 – Topic 2 - SNR_Sample.xlsx LCC_Sample.xlsx

On Class Practice Data – Iris.txt (Neucom format) and your own data (if applicable) Method: PCA, LDA, SNR Software – Neucom v0.919 – Steps: Visualization->PCA – Steps: Visualization->LDA – Steps: Data Analysis->SNR