Mining Binary Constraints in the Construction of Feature Models Li Yi Peking University March 30, 2012.

Slides:



Advertisements
Similar presentations
Yansong Feng and Mirella Lapata
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Sequential Minimal Optimization Advanced Machine Learning Course 2012 Fall Semester Tsinghua University.
Chapter 5: Introduction to Information Retrieval
Introduction to Information Retrieval
An Introduction of Support Vector Machine
Classification / Regression Support Vector Machines
Imbalanced data David Kauchak CS 451 – Fall 2013.
Support Vector Machines Instructor Max Welling ICS273A UCIrvine.
SVM—Support Vector Machines
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
LOGO Classification IV Lecturer: Dr. Bo Yuan
Particle swarm optimization for parameter determination and feature selection of support vector machines Shih-Wei Lin, Kuo-Ching Ying, Shih-Chieh Chen,
Classification of the aesthetic value of images based on histogram features By Xavier Clements & Tristan Penman Supervisors: Vic Ciesielski, Xiadong Li.
A Survey on Text Categorization with Machine Learning Chikayama lab. Dai Saito.
GENERATING AUTOMATIC SEMANTIC ANNOTATIONS FOR RESEARCH DATASETS AYUSH SINGHAL AND JAIDEEP SRIVASTAVA CS DEPT., UNIVERSITY OF MINNESOTA, MN, USA.
Jianwei Lu1 Information Extraction from Event Announcements Student: Jianwei Lu ( ) Supervisor: Robert Dale.
Support Vector Machines
SUPPORT VECTOR MACHINES PRESENTED BY MUTHAPPA. Introduction Support Vector Machines(SVMs) are supervised learning models with associated learning algorithms.
Chapter 7: Text mining UIC - CS 594 Bing Liu 1 1.
Semantic text features from small world graphs Jure Leskovec, IJS + CMU John Shawe-Taylor, Southampton.
Parameterizing Random Test Data According to Equivalence Classes Chris Murphy, Gail Kaiser, Marta Arias Columbia University.
Bing LiuCS Department, UIC1 Learning from Positive and Unlabeled Examples Bing Liu Department of Computer Science University of Illinois at Chicago Joint.
Text Classification With Labeled and Unlabeled Data Presenter: Aleksandar Milisic Supervisor: Dr. David Albrecht.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Introduction to Machine Learning Approach Lecture 5.
Walter Hop Web-shop Order Prediction Using Machine Learning Master’s Thesis Computational Economics.
SVMLight SVMLight is an implementation of Support Vector Machine (SVM) in C. Download source from :
Ch. Eick: Support Vector Machines: The Main Ideas Reading Material Support Vector Machines: 1.Textbook 2. First 3 columns of Smola/Schönkopf article on.
Text Classification using SVM- light DSSI 2008 Jing Jiang.
Li Yi, APSEC ‘12 Constructing Feature Models Us­­ing a Cross-Join Merging Operator.
Prediction model building and feature selection with SVM in breast cancer diagnosis Cheng-Lung Huang, Hung-Chang Liao, Mu- Chen Chen Expert Systems with.
ADVANCED CLASSIFICATION TECHNIQUES David Kauchak CS 159 – Fall 2014.
“Study on Parallel SVM Based on MapReduce” Kuei-Ti Lu 03/12/2015.
Machine Learning CSE 681 CH2 - Supervised Learning.
Improving Web Spam Classification using Rank-time Features September 25, 2008 TaeSeob,Yun KAIST DATABASE & MULTIMEDIA LAB.
GA-Based Feature Selection and Parameter Optimization for Support Vector Machine Cheng-Lung Huang, Chieh-Jen Wang Expert Systems with Applications, Volume.
1 SUPPORT VECTOR MACHINES İsmail GÜNEŞ. 2 What is SVM? A new generation learning system. A new generation learning system. Based on recent advances in.
Active Learning on Spatial Data Christine Körner Fraunhofer AIS, Uni Bonn.
Classification and Ranking Approaches to Discriminative Language Modeling for ASR Erinç Dikici, Murat Semerci, Murat Saraçlar, Ethem Alpaydın 報告者:郝柏翰 2013/01/28.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
ICML2004, Banff, Alberta, Canada Learning Larger Margin Machine Locally and Globally Kaizhu Huang Haiqin Yang, Irwin King, Michael.
Contextual Ranking of Keywords Using Click Data Utku Irmak, Vadim von Brzeski, Reiner Kraft Yahoo! Inc ICDE 09’ Datamining session Summarized.
Xiangnan Kong,Philip S. Yu Multi-Label Feature Selection for Graph Classification Department of Computer Science University of Illinois at Chicago.
Mining Binary Constraints in Feature Models: A Classification-based Approach Yi Li.
Competence Centre on Information Extraction and Image Understanding for Earth Observation 29th March 2007 Category - based Semantic Search Engine 1 Mihai.
Bing LiuCS Department, UIC1 Chapter 8: Semi-supervised learning.
Collaborative Feature Modeling: An Extendable Voting-Based Approach with Divergence Tolerance and Consensus Facilitation Li Yi
ProjFocusedCrawler CS5604 Information Storage and Retrieval, Fall 2012 Virginia Tech December 4, 2012 Mohamed M. G. Farag Mohammed Saquib Khan Prasad Krishnamurthi.
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
Support Vector Machines and Gene Function Prediction Brown et al PNAS. CS 466 Saurabh Sinha.
***Classification Model*** Hosam Al-Samarraie, PhD. CITM-USM.
Chapter 6. Classification and Prediction Classification by decision tree induction Bayesian classification Rule-based classification Classification by.
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
 Effective Multi-Label Active Learning for Text Classification Bishan yang, Juan-Tao Sun, Tengjiao Wang, Zheng Chen KDD’ 09 Supervisor: Koh Jia-Ling Presenter:
Ping-Tsun Chang Intelligent Systems Laboratory NTU/CSIE Using Support Vector Machine for Integrating Catalogs.
Elizabeth R McMahon 14 April 2017
Name: Sushmita Laila Khan Affiliation: Georgia Southern University
Support Vector Machines (SVM)
Damiano Bolzoni, Sandro Etalle, Pieter H. Hartel
An Introduction to Support Vector Machines
Large Scale Support Vector Machines
Applying SVM to Data Bypass Prediction
Shih-Wei Lin, Kuo-Ching Ying, Shih-Chieh Chen, Zne-Jung Lee
Semi-Automatic Data-Driven Ontology Construction System
Physics-guided machine learning for milling stability:
Practice Project Overview
Extracting Why Text Segment from Web Based on Grammar-gram
Presentation transcript:

Mining Binary Constraints in the Construction of Feature Models Li Yi Peking University March 30, 2012

Agenda  Introduction  The Approach  Experiments  Conclusions & Future Work

Background: Feature Models  Feature Model  Construction (Domain Engineering) Requirements Feature Tree + Cross-Tree Constraints  Reuse (Application Engineering) Select a subset of features without violating constraints Audio Playing Software Burn CD Platform PC Mobile Audio CD Codec Optional Mandatory XOR-Group Requires Excludes EXAMPLE: A Feature Model of Audio Playing Software Domain

Help the Construction of FMs Feature Model = Feature Tree + Cross-Tree Constraints The process needs a broad review of requirements documents of existing applications in a domain [1] (Semi-) Automation Supported ? [1] Kang et al. FODA Feasibility Study

Finding Constraints is Challenging  Size of Problem Space: O(|Feature| 2 )  Feature: Often concrete, can be directly observed from an individual product vs. Constraint: Often abstract, have to be learned from a family of similar products  My Experience: Finding constraints is challenging for 30+ features  Real FMs tend to have features We try to provide some automation support.

Our Basic Idea

Agenda  Introduction  The Approach  Experiments  Conclusions & Future Work

Approach Overview Make Feature Pairs Training & Test Feature Models Training & Test Feature Pairs Training Vectors Test Vectors Classified Test Feature Pairs Quantify Feature Pairs TrainOptimize Test Classifier Trained Classifier Feature Pair name 1 : String name 2 : String description 1 : Text description 2 : Text

Agenda  Introduction  Approach: Details  Make & Quantify Feature Pairs  Experiments  Conclusions & Future Work

Make Pairs  The pairs are cross-tree only and unordered  Cross-tree only: The 2 features in a pair have no “ancestor-descendant” relation … Feature Tree A B Y X C (A, B) (A, X) (A, Y) (A, C) (B, X) (B, Y) (B, C) (X, Y) (C, X) (C, Y)  Unordered: (A, B) == (B, A) requires(A, B): A requires B or B requires A or both

Quantify Pairs  We measure 4 numeric attributes for pair (A, B) 1.Similarity between A.description and B.description 2.Similarity between A.objects and B.objects 3.Similarity between A.name and B.objects 4.Similarity between A.objects and B.name Feature Pair name 1 : String name 2 : String description 1 : Text description 2 : Text Feature Pair attribute 1 : Number attribute 2 : Number …(more) Classifiers work with numbers only. Overlapped Function Area Similar Feature One is targeted by another These phenomena may indicate dependency / interaction between the paired features, and in turn, indicate constraints between them.

Extract Objects

Calculate the Similarity tf idf dot-product

Agenda  Introduction  Approach: Details  Train and Optimize the Classifier  Experiments  Conclusions & Future Work

The Classifier: Support Vector Machine (SVM) Idea: Find a separating hyperplane with maximal margin. Implementation: The LIBSVM tool

Optimize the Classifier k = 4 Rationale: Correctly classify a rare class is more important.

Agenda  Introduction  Approach: Details  Experiments  Conclusions & Future Work

Data Preparation  FMs in experiments are built by third parties  From SPLOT Feature Model Repository [1] (no feature description)  Graph Product Line: by Don Batory (91 pairs)  Weather Station: by pure-systems corp. (196 pairs)  Add Feature Description  Most features = Domain terminologies Search terms in Wikipedia Description = The first paragraph (i.e. the Abstract)  Other features: No description [1]

Experiments Design Generate Training & Test Set Optimize, Train and Test Results No Feedback Generate Initial Training & Test Set Optimize, Train and Test Results Training & Test Set Check a few results Add checked results to training set; Remove them from test set Add checked results to training set; Remove them from test set Limited Feedback (An expected practice in real world)  3 Training / Test Set Selection Strategies  Cross-Domain: Training = FM 1, Test = FM 2  Inner-Domain: Training = 1/5 of FM 2, Test = Rest of FM 2  Hybrid: Training = FM 1 + 1/5 of FM 2, Test = Rest of FM 2  2 FMs: one as FM 1, another as FM 2 ; then exchange  2 Training Methods  Normal: Training with known data (i.e. training set)  LU-Method: Iterated training with known and unknown data  With or Without (Limited) Feedback

Measurements Predicted PositivePredicted Negative Actual PositiveTrue Positive (TP)False Negative (FN) Actual NegativeFalse Positive (FP)True Negative (TN)

Results: Optimization Feature Model Training = WS, Test = GPLTraining = GPL, Test = WS Strategy Cross- Domain Inner- Domain Hybrid Cross- Domain Inner- Domain Hybrid Avg. Error % (with Default Parameter Values) Avg. Error % (Optimized) Before: Unstable (3% ~ 73%) After: Stable (1% ~ 13%) The optimization results are very similar to those reported in general classification research papers.

Results: Without Feedback Requires Excludes Precision % Recall % F 2 -MeasurePrecision %Recall %F 2 -Measure LLUL L L L L Training FM = Weather Station, Test FM = Graph Product Line Cross- Domain N/A 00 Inner- Domain Hybrid Training FM = Graph Product Line, Test FM = Weather Station Cross- Domain N/A 00 Inner- Domain Hybrid L = Normal Training, LU = LU-Training  The cross-domain strategy fails to find any excludes.  No significant difference between inner-domain and hybrid strategies.  Recall is high. Precision depends on the test FM (unstable).  No significant difference between normal and LU- training, so we prefer the former one for saving training time.

Results: Normal Training + Feedback 3 feedbacks/ turn (i.e. 2% ~ 5% data) 10 turns Improve Recall. Precision is still fluctuate. Help cross-domain find excludes.

Agenda  Introduction  Approach: Details  Experiments  Conclusions & Future Work

Conclusions & Future Work  Conclusions  Binary constraints between features  Classes of feature- pairs  The classifier should be optimized  High recall  Unstable precision  Preferred Settings: Inner-Domain/Hybrid Training Set + Normal Training + Limited Feedback  Future Work  More linguistic analysis (verb, time, etc.)  Real use

THANK YOU ! Q&A