Feature Selection 1 Feature Selection for Image Retrieval By Karina Zapién Arreola January 21th, 2005.

Slides:



Advertisements
Similar presentations
Numbers Treasure Hunt Following each question, click on the answer. If correct, the next page will load with a graphic first – these can be used to check.
Advertisements

Scenario: EOT/EOT-R/COT Resident admitted March 10th Admitted for PT and OT following knee replacement for patient with CHF, COPD, shortness of breath.
Simplifications of Context-Free Grammars
Angstrom Care 培苗社 Quadratic Equation II
AP STUDY SESSION 2.
1
Copyright © 2003 Pearson Education, Inc. Slide 1 Computer Systems Organization & Architecture Chapters 8-12 John D. Carpinelli.
Copyright © 2011, Elsevier Inc. All rights reserved. Chapter 6 Author: Julia Richards and R. Scott Hawley.
Author: Julia Richards and R. Scott Hawley
Myra Shields Training Manager Introduction to OvidSP.
STATISTICS Joint and Conditional Distributions
Properties Use, share, or modify this drill on mathematic properties. There is too much material for a single class, so you’ll have to select for your.
David Burdett May 11, 2004 Package Binding for WS CDL.
1 RA I Sub-Regional Training Seminar on CLIMAT&CLIMAT TEMP Reporting Casablanca, Morocco, 20 – 22 December 2005 Status of observing programmes in RA I.
Properties of Real Numbers CommutativeAssociativeDistributive Identity + × Inverse + ×
Local Customization Chapter 2. Local Customization 2-2 Objectives Customization Considerations Types of Data Elements Location for Locally Defined Data.
Process a Customer Chapter 2. Process a Customer 2-2 Objectives Understand what defines a Customer Learn how to check for an existing Customer Learn how.
Human Service Providers and Referrals Chapter 5. Human Service Providers and Referrals 5-2 Objectives Demonstrate the process for entering a Human Service.
Custom Statutory Programs Chapter 3. Customary Statutory Programs and Titles 3-2 Objectives Add Local Statutory Programs Create Customer Application For.
Custom Services and Training Provider Details Chapter 4.
CALENDAR.
1 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt BlendsDigraphsShort.
Chapter 7 Sampling and Sampling Distributions
1 Click here to End Presentation Software: Installation and Updates Internet Download CD release NACIS Updates.
Solve Multi-step Equations
1 Combination Symbols A supplement to Greenleafs QR Text Compiled by Samuel Marateck ©2009.
Break Time Remaining 10:00.
Factoring Quadratics — ax² + bx + c Topic
Turing Machines.
Table 12.1: Cash Flows to a Cash and Carry Trading Strategy.
McGraw-Hill/Irwin McGraw-Hill/Irwin Copyright © 2009 by The McGraw-Hill Companies, Inc. All rights reserved.
PP Test Review Sections 6-1 to 6-6
1 The Blue Café by Chris Rea My world is miles of endless roads.
Bright Futures Guidelines Priorities and Screening Tables
EIS Bridge Tool and Staging Tables September 1, 2009 Instructor: Way Poteat Slide: 1.
Bellwork Do the following problem on a ½ sheet of paper and turn in.
CS 6143 COMPUTER ARCHITECTURE II SPRING 2014 ACM Principles and Practice of Parallel Programming, PPoPP, 2006 Panel Presentations Parallel Processing is.
Operating Systems Operating Systems - Winter 2010 Chapter 3 – Input/Output Vrije Universiteit Amsterdam.
Exarte Bezoek aan de Mediacampus Bachelor in de grafische en digitale media April 2014.
Copyright © 2013, 2009, 2006 Pearson Education, Inc. 1 Section 5.5 Dividing Polynomials Copyright © 2013, 2009, 2006 Pearson Education, Inc. 1.
Copyright © 2012, Elsevier Inc. All rights Reserved. 1 Chapter 7 Modeling Structure with Blocks.
1 RA III - Regional Training Seminar on CLIMAT&CLIMAT TEMP Reporting Buenos Aires, Argentina, 25 – 27 October 2006 Status of observing programmes in RA.
Basel-ICU-Journal Challenge18/20/ Basel-ICU-Journal Challenge8/20/2014.
1..
CONTROL VISION Set-up. Step 1 Step 2 Step 3 Step 5 Step 4.
Adding Up In Chunks.
MaK_Full ahead loaded 1 Alarm Page Directory (F11)
1 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt Synthetic.
1 Using Bayesian Network for combining classifiers Leonardo Nogueira Matos Departamento de Computação Universidade Federal de Sergipe.
Subtraction: Adding UP
: 3 00.
5 minutes.
1 hi at no doifpi me be go we of at be do go hi if me no of pi we Inorder Traversal Inorder traversal. n Visit the left subtree. n Visit the node. n Visit.
Analyzing Genes and Genomes
Speak Up for Safety Dr. Susan Strauss Harassment & Bullying Consultant November 9, 2012.
Essential Cell Biology
Converting a Fraction to %
Chapter 8 Estimation Understandable Statistics Ninth Edition
Exponents and Radicals
Clock will move after 1 minute
PSSA Preparation.
Essential Cell Biology
Immunobiology: The Immune System in Health & Disease Sixth Edition
Simple Linear Regression Analysis
Physics for Scientists & Engineers, 3rd Edition
Energy Generation in Mitochondria and Chlorplasts
Select a time to count down from the clock above
1 Dr. Scott Schaefer Least Squares Curves, Rational Representations, Splines and Continuity.
1 Decidability continued…. 2 Theorem: For a recursively enumerable language it is undecidable to determine whether is finite Proof: We will reduce the.
Presentation transcript:

Feature Selection 1 Feature Selection for Image Retrieval By Karina Zapién Arreola January 21th, 2005

Feature Selection 2 Introduction Variable and feature selection have become the focus of much research in areas of applications for datasets with many variables are available Text processing Gene expression Combinatorial chemistry

Feature Selection 3 Motivation The objective of feature selection is three- fold: Improving the prediction performance of the predictors Providing a faster and more cost-effective predictors Providing a better understanding of the underlying process that generated the data

Feature Selection 4 Why use feature selection in CBIR Different users may need different features for image retrieval From each selected sample, a specific feature set can be chosen

Feature Selection 5 Boosting Method for improving the accuracy of any learning algorithm Use of “weak algorithms” for single rules Weighting of the weak algorithms Combination of weak rules into a strong learning algorithm

Feature Selection 6 Adaboost Algorithm Is a iterative boosting algorithm Notation Samples (x 1,y 1 ),…,(x m,y m ), where, y i = -1,1 There are m positive samples, and l negative samples Weak classifiers h i For iteration t, the error is defined as: ε t = min (½)Σ i ω i |h i (x i ) – y i | where ω i is a weight for x i.

Feature Selection 7 Adaboost Algorithm Given samples (x 1,y 1 ),…,(x m,y m ), where y i = -1,1 Initialize ω 1,i =1/(2m), 1/(2l), for y i = 1,-1 For t=1,…,T Normalize ω t,i = ω t,i /(Σ j ω t,j ) Train base learner h t,i using distribution ω i,j Choose h t that minimize ε t with error e i Update ω t+1,i = ω t,i β t 1-e i Set β t = (ε t )/(1- ε t ) and α t = log(1/ β t ) Output the final classifier H(x) = sign( Σ t α t h t (x) )

Feature Selection 8 Adaboost Application Searching similar groups A particular image class is chosen A positive sample of this group is given randomly A negative sample of the rest of the images is given randomly

Feature Selection 9 Check list Feature Selection Domain knowledge Commensurate features Interdependence of features Prune of input variables Asses features individually Dirty data Predictor – linear predictor Comparison Stable solution

Feature Selection 10 Domain knowledge Features used colordb_sumRGB _entropy_d1 col_gpd_hsv col_gpd_lab col_gpd_rgb col_hu_hsv2 col_hu_lab2 col_hu_lab col_hu_rgb2 col_hu_rgb col_hu_seg2_hsv col_hu_seg2_lab col_hu_seg2_rgb Features used col_hu_seg_hsv col_hu_seg_lab col_hu_seg_rgb col_hu_yiq col_ngcm_rgb col_sm_hsv col_sm_lab col_sm_rgb col_sm_yiq text_gabor text_tamura edgeDB waveletDB Features used hist_phc_hsv hist_phc_rgb Hist_Grad_RGB haar_RGB haar_HSV haar_rgb haar_hmmd

Feature Selection 11 Check list Feature Selection Domain knowledge Commensurate features Normalize features between an appropriated range Adaboost takes each feature independent so it is not necessary to normalize them

Feature Selection 12 Check list Feature Selection Domain knowledge Commensurate features Interdependence of features Prune of input variables Asses features individually Dirty data Predictor – linear predictor Comparison Stable solution

Feature Selection 13 Feature construction and space dimensionality reduction Clustering Correlation coefficient Supervised feature selection Filters

Feature Selection 14 Check list Feature Selection Domain knowledge Commensurate features Interdependence of features Prune of input variables Features with the same value for all samples (variance=0) were eliminated From 4912 Linear Features 3583 were selected

Feature Selection 15 Check list Feature Selection Domain knowledge Commensurate features Interdependence of features Prune of input variables Asses features individually When there is no asses method, use Variable Ranking method. In Adaboost this is not necessary

Feature Selection 16 Variable Ranking Preprocessing step Independent of the choice of the predictor Correlation criteria It can only detect linear dependencies Single variable classifiers

Feature Selection 17 Variable Ranking Noise reduction and better classification may be obtained by adding variables that are presumable redundant Perfectly correlated variables are truly redundant in the sense that no additional information is gained by adding them. It doesn’t mean absence of variable complementarily Two variables that are useless by themselves can be useful together

Feature Selection 18 Check list Feature Selection Domain knowledge Commensurate features Interdependence of features Prune of input variables Asses features individually Dirty data Predictor – linear predictor Comparison Stable solution

Feature Selection 19 Check list Feature Selection Domain knowledge Commensurate features Interdependence of features Prune of input variables Asses features individually Dirty data Predictor – linear predictor Comparison Stable solution

Feature Selection 20 Adaboost Algorithm Given samples (x 1,y 1 ),…,(x m,y m ), where x i, y i -1,1 Initialize ω 1,i =1/(2m), 1/(2l), for y i = -1,1 For t=1,…,T Normalize ω t,i = ω t,i /(Σ j ω t,j ) Train base learner h t,i using distribution ω i,j Choose h t that minimize ε t with error e i Update ω t+1,i = ω t,i β t 1-e i Set β t = (ε t )/(1- ε t ) and α t = log(1/ β t ) Output the final classifier H(x) = sign( Σ t α t h t (x) )

Feature Selection 21 Weak classifier Each weak classifier h i is defined as follows: h i.pos_mean – mean value for positive samples h i.neg_mean – mean value for negative sample A sample is classified as: 1 if it is closer to h i.pos_mean -1 if it is closer to h i.neg_mean

Feature Selection 22 Weak classifier h i.pos_mean – mean value for positive samples h i.neg_mean – mean value for negative sample A Linear Classifier was used h i.neg_meanh i.pos_mean

Feature Selection 23 Check list Feature Selection Domain knowledge Commensurate features Interdependence of features Prune of input variables Asses features individually Dirty data Predictor – linear predictor Comparison Stable solution

Feature Selection 24 Adaboost experiments and results 4 positives 10 positives

Feature Selection 25 Few positive samples Use of 4 positive samples

Feature Selection 26 More positive samples False Positive Use of 10 positive samples

Feature Selection 27 Training data Test data False negative Use of 10 positive samples

Feature Selection 28 Changing number of Training Iterations The number of iterations Used was from 5 to 50 Iterations = 30 was set

Feature Selection 29 Changing Sample Size 5 pos10 pos 15 pos 20 pos25 pos 30 pos 35 pos

Feature Selection 30 Few negative samples Use of 15 negative samples

Feature Selection 31 More negative samples Use of 75 negative samples

Feature Selection 32 Check list Feature Selection Domain knowledge Commensurate features Interdependence of features Prune of input variables Asses features individually Dirty data Predictor – linear predictor Comparison (ideas, time, comp. resources, examples) Stable solution

Feature Selection 33 Stable solution For Adaboost is important to have a representative sample Chosen parameters: Positives samples: 15 Negative samples: 100 Iteration number: 30

Feature Selection 34 Stable solution with more samples and iterations Beaches Dinosaurs Mountains Elephants Buildings Humans Roses Buses Horses Food

Feature Selection 35 Stable solution for Dinosaurs Use of: 15 Positive samples 100 Negative samples 30 Iterations

Feature Selection 36 Stable solution for Roses Use of: 15 Positive samples 100 Negative samples 30 Iterations

Feature Selection 37 Stable solution for Buses Use of: 15 Positive samples 100 Negative samples 30 Iterations

Feature Selection 38 Stable solution for Beaches Use of: 15 Positive samples 100 Negative samples 30 Iterations

Feature Selection 39 Stable solution for Food Use of: 15 Positive samples 100 Negative samples 30 Iterations

Feature Selection 40 Unstable Solution

Feature Selection 41 Unstable solution for Roses Use of: 5 Positive samples 10 Negative samples 30 Iterations

Feature Selection 42 Best features for classification Humans Beaches Buildings Buses Dinosaurs Elephants Roses Horses Mountains Food

Feature Selection 43 And the winner is…

Feature Selection 44 Feature frequency

Feature Selection 45 Extensions Searching similar images Pairs of images are built The difference for each feature is calculated Each difference is classified as: 1 if both images belong to the same class -1 if both images belong to different classes Multiclass adaboost

Feature Selection 46 Extensions Use of another weak classifier Design weak classifier using multiple features → classifier fusion Use different weak classifier such as SVM, NN, threshold function, etc. Different feature selection method: SVM

Feature Selection 47 Discussion Is important to add feature Selection for Image retrieval A good methodology for selecting features should be used Adaboost is a learning algorithm → data dependent It is important to have representative samples Adaboost can help to improve the classification potential of simple algorithms

Feature Selection 48 Thank you !