Shih-Wei Lin, Kuo-Ching Ying, Shih-Chieh Chen, Zne-Jung Lee

Slides:



Advertisements
Similar presentations
(SubLoc) Support vector machine approach for protein subcelluar localization prediction (SubLoc) Kim Hye Jin Intelligent Multimedia Lab
Advertisements

Generative Models Thus far we have essentially considered techniques that perform classification indirectly by modeling the training data, optimizing.
Application a hybrid controller to a mobile robot J.-S Chiou, K. -Y. Wang,Simulation Modelling Pratice and Theory Vol. 16 pp (2008) Professor:
Ali Husseinzadeh Kashan Spring 2010
ECG Signal processing (2)
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
SVM - Support Vector Machines A new classification method for both linear and nonlinear data It uses a nonlinear mapping to transform the original training.
An Introduction of Support Vector Machine
An Introduction of Support Vector Machine
Support Vector Machines
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Particle swarm optimization for parameter determination and feature selection of support vector machines Shih-Wei Lin, Kuo-Ching Ying, Shih-Chieh Chen,
Particle Swarm Optimization (PSO)  Kennedy, J., Eberhart, R. C. (1995). Particle swarm optimization. Proc. IEEE International Conference.
Face Recognition & Biometric Systems Support Vector Machines (part 2)
The Disputed Federalist Papers : SVM Feature Selection via Concave Minimization Glenn Fung and Olvi L. Mangasarian CSNA 2002 June 13-16, 2002 Madison,
SUPPORT VECTOR MACHINES PRESENTED BY MUTHAPPA. Introduction Support Vector Machines(SVMs) are supervised learning models with associated learning algorithms.
Optimizing F-Measure with Support Vector Machines David R. Musicant Vipin Kumar Aysel Ozgur FLAIRS 2003 Tuesday, May 13, 2003 Carleton College.
Attention Deficit Hyperactivity Disorder (ADHD) Student Classification Using Genetic Algorithm and Artificial Neural Network S. Yenaeng 1, S. Saelee 2.
Identifying Computer Graphics Using HSV Model And Statistical Moments Of Characteristic Functions Xiao Cai, Yuewen Wang.
Efficient Model Selection for Support Vector Machines
Prediction model building and feature selection with SVM in breast cancer diagnosis Cheng-Lung Huang, Hung-Chang Liao, Mu- Chen Chen Expert Systems with.
Feature Selection in Nonlinear Kernel Classification Olvi Mangasarian & Edward Wild University of Wisconsin Madison Workshop on Optimization-Based Data.
Feature Selection in Nonlinear Kernel Classification Olvi Mangasarian Edward Wild University of Wisconsin Madison.
Fuzzy Entropy based feature selection for classification of hyperspectral data Mahesh Pal Department of Civil Engineering National Institute of Technology.
A Fast Clustering-Based Feature Subset Selection Algorithm for High- Dimensional Data.
GA-Based Feature Selection and Parameter Optimization for Support Vector Machine Cheng-Lung Huang, Chieh-Jen Wang Expert Systems with Applications, Volume.
GATree: Genetically Evolved Decision Trees 전자전기컴퓨터공학과 데이터베이스 연구실 G 김태종.
1 SUPPORT VECTOR MACHINES İsmail GÜNEŞ. 2 What is SVM? A new generation learning system. A new generation learning system. Based on recent advances in.
A Regression Approach to Music Emotion Recognition Yi-Hsuan Yang, Yu-Ching Lin, Ya-Fan Su, and Homer H. Chen, Fellow, IEEE IEEE TRANSACTIONS ON AUDIO,
Machine Learning Using Support Vector Machines (Paper Review) Presented to: Prof. Dr. Mohamed Batouche Prepared By: Asma B. Al-Saleh Amani A. Al-Ajlan.
C OMBINING E NSEMBLE T ECHNIQUE OF S UPPORT V ECTOR M ACHINES WITH THE O PTIMAL K ERNEL M ETHOD FOR H IGH D IMENSIONAL D ATA C LASSIFICATION I-Ling Chen.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
Proximal Support Vector Machine Classifiers KDD 2001 San Francisco August 26-29, 2001 Glenn Fung & Olvi Mangasarian Data Mining Institute University of.
A hybrid SOFM-SVR with a filter-based feature selection for stock market forecasting Huang, C. L. & Tsai, C. Y. Expert Systems with Applications 2008.
An Introduction to Support Vector Machines (M. Law)
Max-Margin Classification of Data with Absent Features Presented by Chunping Wang Machine Learning Group, Duke University July 3, 2008 by Chechik, Heitz,
Exploring Alternative Splicing Features using Support Vector Machines Feature for Alternative Splicing Alternative splicing is a mechanism for generating.
A Simulated-annealing-based Approach for Simultaneous Parameter Optimization and Feature Selection of Back-Propagation Networks (BPN) Shih-Wei Lin, Tsung-Yuan.
RSVM: Reduced Support Vector Machines Y.-J. Lee & O. L. Mangasarian First SIAM International Conference on Data Mining Chicago, April 6, 2001 University.
컴퓨터 과학부 김명재.  Introduction  Data Preprocessing  Model Selection  Experiments.
Computational Approaches for Biomarker Discovery SubbaLakshmiswetha Patchamatla.
Improving Support Vector Machine through Parameter Optimized Rujiang Bai, Junhua Liao Shandong University of Technology Library Zibo , China { brj,
Proximal Plane Classification KDD 2001 San Francisco August 26-29, 2001 Glenn Fung & Olvi Mangasarian Second Annual Review June 1, 2001 Data Mining Institute.
Nawanol Theera-Ampornpunt, Seong Gon Kim, Asish Ghoshal, Saurabh Bagchi, Ananth Grama, and Somali Chaterji Fast Training on Large Genomics Data using Distributed.
Machine Learning and Data Mining: A Math Programming- Based Approach Glenn Fung CS412 April 10, 2003 Madison, Wisconsin.
Next, this study employed SVM to classify the emotion label for each EEG segment. The basic idea is to project input data onto a higher dimensional feature.
A distributed PSO – SVM hybrid system with feature selection and parameter optimization Cheng-Lung Huang & Jian-Fan Dun Soft Computing 2008.
A Brief Introduction to Support Vector Machine (SVM) Most slides were from Prof. A. W. Moore, School of Computer Science, Carnegie Mellon University.
Hybrid Ant Colony Optimization-Support Vector Machine using Weighted Ranking for Feature Selection and Classification.
CS 9633 Machine Learning Support Vector Machines
PREDICT 422: Practical Machine Learning
Particle Swarm Optimization (2)
Zhenshan, Wen SVM Implementation Zhenshan, Wen
Particle Swarm Optimization with Partial Search To Solve TSP
Rule Induction for Classification Using
Particle Swarm Optimization
PSO -Introduction Proposed by James Kennedy & Russell Eberhart in 1995
Meta-heuristics Introduction - Fabien Tricoire
Glenn Fung, Murat Dundar, Bharat Rao and Jinbo Bi
Project 4: Facial Image Analysis with Support Vector Machines
An Enhanced Support Vector Machine Model for Intrusion Detection
Avdesh Mishra, Manisha Panta, Md Tamjidul Hoque, Joel Atallah
Advanced Artificial Intelligence Evolutionary Search Algorithm
An Introduction to Support Vector Machines
Pattern Recognition CS479/679 Pattern Recognition Dr. George Bebis
Discriminative Frequent Pattern Analysis for Effective Classification
Concave Minimization for Support Vector Machine Classifiers
A Novel Smoke Detection Method Using Support Vector Machine
University of Wisconsin - Madison
Presentation transcript:

Shih-Wei Lin, Kuo-Ching Ying, Shih-Chieh Chen, Zne-Jung Lee Particle swarm optimization for parameter determination and feature selection of support vector machines Shih-Wei Lin, Kuo-Ching Ying, Shih-Chieh Chen, Zne-Jung Lee Expert Systems with Applications 2008

Introduction Classification problems have been extensively studied. Support vector machine (SVM) is a popular pattern classification method with many diverse applications. Kernel parameter setting in the SVM training procedure, along with the feature selection, significantly influences the classification accuracy.

Introduction How to choose the optimal input feature subset for SVM? How to set the best kernel parameters? Hybridizing the particle swarm optimization (PSO) and SVM to improve the classification accuracy with a small and appropriate feature subset. This makes the optimal separating hyper-plane obtainable in both linear and non-linear classification problems 1. 2.

Support Vector Machine (SVM) Support vector machine (SVM) is a new technique for data classification were first suggested by Vapnik in 1995. SVM is using Separating Hyperplane to distinguish the data of two or several different Class that deal with the data mining problem of classification.

Kernel Function Several kernel functions help the SVM in obtaining the optimal solution. For example: Linear, Polynomial, RBF, Sigmoid The RBF is generally applied most frequently, because it can classify multi-dimensional data, unlike a linear kernel function. The RBF has fewer parameters to set than other kernel. RBF is an effective option for kernel function.

SVM Parameters Setting Proper parameters setting can improve the classification accuracy of SVM. The parameters that should be optimized include penalty parameter C and the parameters with different kernel function (gamma of RBF). Grid Algorithm is an alternative to find the best C and the gamma parameter, however it is time consuming and does not perform well.

Feature Selection Feature selection is used to identify a powerfully predictive subset of fields within the database and to reduce the number of fields presented to the mining process. Affects several aspects of pattern classification: 1.The accuracy of classification algorithm learned 2.The time needed for learning a classification function 3.The number of examples needed for learning 4.The cost associated with feature

Feature Selection Filter Wrapper

Particle swarm optimization Particle swarm optimization (PSO) (Kennedy & Eberhart,1995) is an emerging population-based meta-heuristic The new position of a particle is calculated using the following formula:

Particle swarm optimization rnd( ) is a random function in the range[0, 1] Positive constant c1 and c2 are personal and social learning factors. w is the inertia weight and inertia weight balances the global exploration and local exploitation. Pi,d denote the best previous position encountered by the ith particle. Pg,d denotes the global best position thus far. t denotes the iteration counter.

Search concept of PSO.

Grid-Search Algorithm

Particle representation

The flowchart of PSO algorithm. Fitness = Accuracy

Platform The platform adopted to develop the PSO + SVM approach is a PC with the following features: Intel Pentium IV 3.0 GHz CPU 512 MB RAM Windows XP operating system Visual C++ 6.0 development environment

Dataset

Cross-Validation To guarantee that the present results are valid and can be generalized for making predictions regarding new data Using k-fold-cross-validation This study used k = 10, meaning that all of the data will be divided into ten parts, each of which will take turns at being the testing data set.

PSO-based parameters determination and feature selection approach for SVM

Comparison between the PSO + SVM, NSVM, SVM, LSVM Fung & Mangasarian (2003) and Liao et al (2004)

PSO+SVM VS PSO+GA

PSO + SVM approach with and without feature selection and grid search

PSO + SVM approach with and without feature selection

Conclusions We proposed a PSO-based strategy to select features subset and to set the parameters for SVM classification. We have conducted two experiments to evaluate the classification accuracy of the proposed PSO-based approach with RBF kernel and the grid search method on 17 real-world datasets from UCI database. Generally, compared with the grid search approach, the proposed PSO-based approach has good accuracy performance with fewer features.

Future research Other kernel parameters can also be optimized using the same approach. Experimental results obtained from UCI datasets, other public datasets and real-world problems can be tested in the future to verify and extend this approach.

Thank You Q & A