ACTIVE LEARNING USING CONFORMAL PREDICTORS: APPLICATION TO IMAGE CLASSIFICATION HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Active Learning with Feedback on Both Features and Instances H. Raghavan, O. Madani and R. Jones Journal of Machine Learning Research 7 (2006) Presented.
A New Evolving Tree for Text Document Clustering and Visualization 1 Wui Lee Chang, 1* Kai Meng Tay, 2 Chee Peng Lim 1 Faculty of Engineering, Universiti.
Document Summarization using Conditional Random Fields Dou Shen, Jian-Tao Sun, Hua Li, Qiang Yang, Zheng Chen IJCAI 2007 Hao-Chin Chang Department of Computer.
Random Forest Predrag Radenković 3237/10
+ Multi-label Classification using Adaptive Neighborhoods Tanwistha Saha, Huzefa Rangwala and Carlotta Domeniconi Department of Computer Science George.
Multimedia Answer Generation for Community Question Answering.
Texture Segmentation Based on Voting of Blocks, Bayesian Flooding and Region Merging C. Panagiotakis (1), I. Grinias (2) and G. Tziritas (3)
Nov 4, Detection, Classification and Tracking of Targets in Distributed Sensor Networks Presented by: Prabal Dutta Dan Li, Kerry Wong,
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
1 Integrating User Feedback Log into Relevance Feedback by Coupled SVM for Content-Based Image Retrieval 9-April, 2005 Steven C. H. Hoi *, Michael R. Lyu.
Active Learning with Support Vector Machines
Multivariate Regression. Bayesian Multivariate Regression.
CSSE463: Image Recognition Day 31 Due tomorrow night – Project plan Due tomorrow night – Project plan Evidence that you’ve tried something and what specifically.
Dept. of Computer Science & Engineering, CUHK Pseudo Relevance Feedback with Biased Support Vector Machine in Multimedia Retrieval Steven C.H. Hoi 14-Oct,
1 Ensembles of Nearest Neighbor Forecasts Dragomir Yankov, Eamonn Keogh Dept. of Computer Science & Eng. University of California Riverside Dennis DeCoste.
Active Learning Strategies for Compound Screening Megon Walker 1 and Simon Kasif 1,2 1 Bioinformatics Program, Boston University 2 Department of Biomedical.
© Prentice Hall1 DATA MINING Introductory and Advanced Topics Part II Margaret H. Dunham Department of Computer Science and Engineering Southern Methodist.
Rotation Forest: A New Classifier Ensemble Method 交通大學 電子所 蕭晴駿 Juan J. Rodríguez and Ludmila I. Kuncheva.
Remote Sensing Laboratory Dept. of Information Engineering and Computer Science University of Trento Via Sommarive, 14, I Povo, Trento, Italy Remote.
1 Efficiently Learning the Accuracy of Labeling Sources for Selective Sampling by Pinar Donmez, Jaime Carbonell, Jeff Schneider School of Computer Science,
Language Identification of Search Engine Queries Hakan Ceylan Yookyung Kim Department of Computer Science Yahoo! Inc. University of North Texas 2821 Mission.
An Approach to Test Autonomic Containers Ronald Stevens (IEEE Computer Society & ACM Student Member) August 1, 2006 REU Sponsored by NSF.
Citation Recommendation 1 Web Technology Laboratory Ferdowsi University of Mashhad.
CSSE463: Image Recognition Day 27 This week This week Last night: k-means lab due. Last night: k-means lab due. Today: Classification by “boosting” Today:
Personalized Information Retrieval in Context David Vallet Universidad Autónoma de Madrid, Escuela Politécnica Superior,Spain.
Towards Minimizing the Annotation Cost of Certified Text Classification Mossaab Bagdouri 1 David D. Lewis 2 William Webber 1 Douglas W. Oard 1 1 University.
Intelligent Database Systems Lab Advisor : Dr. Hsu Graduate : Chien-Shing Chen Author : Satoshi Oyama Takashi Kokubo Toru lshida 國立雲林科技大學 National Yunlin.
Classification and Ranking Approaches to Discriminative Language Modeling for ASR Erinç Dikici, Murat Semerci, Murat Saraçlar, Ethem Alpaydın 報告者:郝柏翰 2013/01/28.
Garrett Poppe, Liv Nguekap, Adrian Mirabel CSUDH, Computer Science Department.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
Xiangnan Kong,Philip S. Yu Multi-Label Feature Selection for Graph Classification Department of Computer Science University of Illinois at Chicago.
Greedy is not Enough: An Efficient Batch Mode Active Learning Algorithm Chen, Yi-wen( 陳憶文 ) Graduate Institute of Computer Science & Information Engineering.
Relative Hidden Markov Models Qiang Zhang, Baoxin Li Arizona State University.
Wikipedia as Sense Inventory to Improve Diversity in Web Search Results Celina SantamariaJulio GonzaloJavier Artiles nlp.uned.es UNED,c/Juan del Rosal,
PSEUDO-RELEVANCE FEEDBACK FOR MULTIMEDIA RETRIEVAL Seo Seok Jun.
21/11/20151Gianluca Demartini Ranking Clusters for Web Search Gianluca Demartini Paul–Alexandru Chirita Ingo Brunkhorst Wolfgang Nejdl L3S Info Lunch Hannover,
CIKM Opinion Retrieval from Blogs Wei Zhang 1 Clement Yu 1 Weiyi Meng 2 1 Department of.
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
AUTOMATIC TARGET RECOGNITION AND DATA FUSION March 9 th, 2004 Bala Lakshminarayanan.
Answering Top-k Queries Using Views Gautam Das (Univ. of Texas), Dimitrios Gunopulos (Univ. of California Riverside), Nick Koudas (Univ. of Toronto), Dimitris.
Active learning Haidong Shi, Nanyi Zeng Nov,12,2008.
CSSE463: Image Recognition Day 33 This week This week Today: Classification by “boosting” Today: Classification by “boosting” Yoav Freund and Robert Schapire.
Multiple Instance Learning for Sparse Positive Bags Razvan C. Bunescu Machine Learning Group Department of Computer Sciences University of Texas at Austin.
Text Document Categorization by Term Association Maria-luiza Antonie Osmar R. Zaiane University of Alberta, Canada 2002 IEEE International Conference on.
Date: 2012/08/21 Source: Zhong Zeng, Zhifeng Bao, Tok Wang Ling, Mong Li Lee (KEYS’12) Speaker: Er-Gang Liu Advisor: Dr. Jia-ling Koh 1.
Iterative similarity based adaptation technique for Cross Domain text classification Under: Prof. Amitabha Mukherjee By: Narendra Roy Roll no: Group:
Ping-Tsun Chang Intelligent Systems Laboratory Computer Science and Information Engineering National Taiwan University Combining Unsupervised Feature Selection.
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
Combining Evolutionary Information Extracted From Frequency Profiles With Sequence-based Kernels For Protein Remote Homology Detection Name: ZhuFangzhi.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
 Effective Multi-Label Active Learning for Text Classification Bishan yang, Juan-Tao Sun, Tengjiao Wang, Zheng Chen KDD’ 09 Supervisor: Koh Jia-Ling Presenter:
Incremental Reduced Support Vector Machines Yuh-Jye Lee, Hung-Yi Lo and Su-Yun Huang National Taiwan University of Science and Technology and Institute.
Mining Concept-Drifting Data Streams Using Ensemble Classifiers Haixun Wang Wei Fan Philip S. YU Jiawei Han Proc. 9 th ACM SIGKDD Internal Conf. Knowledge.
Active, Semi-Supervised Learning for Textual Information Access Anastasia Krithara¹, Cyril Goutte², Massih-Reza Amini³, Jean-Michel Renders¹ Massih-Reza.
What is an Object? —— an experimental evaluation Presented by: Yao Pan.
11 Automated multi-label text categorization with VG-RAM weightless neural networks Presenter: Guan-Yu Chen A. F. DeSouza, F. Pedroni, E. Oliveira, P.
A Pool of Deep Models for Event Recognition
Source: Procedia Computer Science(2015)70:
Active Learning Intrusion Detection using k-Means Clustering Selection
Basic machine learning background with Python scikit-learn
A New Approach to Track Multiple Vehicles With the Combination of Robust Detection and Two Classifiers Weidong Min , Mengdan Fan, Xiaoguang Guo, and Qing.
Finding Clusters within a Class to Improve Classification Accuracy
A Hybrid PCA-LDA Model for Dimension Reduction Nan Zhao1, Washington Mio2 and Xiuwen Liu1 1Department of Computer Science, 2Department of Mathematics Florida.
Active learning The learning algorithm must have some control over the data from which it learns It must be able to query an oracle, requesting for labels.
How To Extend the Training Data
Using Association Rules as Texture features
Presentation transcript:

ACTIVE LEARNING USING CONFORMAL PREDICTORS: APPLICATION TO IMAGE CLASSIFICATION HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions L. Makili 1, J. Vega 2, S. Dormido-Canto 3 1 Universidade Katyavala Bwila. Benguela (Angola) 2 Asociación EURATOM/CIEMAT para Fusión. Madrid (Spain) 3 Universidad Nacional de Educación a Distancia (UNED). Madrid (Spain) 7th Workshop on Fusion Data Processing Validation and Analysis

Outline Introduction Concepts overview Experimental results Conclusions HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions

Motivation 5 – class classification problem – Classification of TJ – II Thomson Scattering images Classifier based on conformal predictors, using SVM as the underlying algorithm – Computationally intensive task Patterns of TSD images: (a) BKGND, (b) COFF, (c) ECRH, (d) NBI and (e) STRAY HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions

Goal To find out a minimal and good enough training dataset for classification purposes HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions

Active learning The learning algorithm must have some control over the data from which it learns It must be able to query an oracle, requesting for labels of data samples that seem to be most informative for the learning process Proper selection of samples implies better performances with fewer data HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions Settles, B. “Active Learning Literature Survey. Computer Sciences Technical Report 1648”, University of Wisconsin – Madison, Available at reports/2009/TR1648.pdfhttp://research.cs.wisc.edu/tech reports/2009/TR1648.pdf

Uncertainty sampling The learning algorithm selects new examples when their class membership is unclear Suitable for classifiers that besides making classification decisions, estimates certainty of these decisions HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions Lewis, D. and Gale, W., “A Sequential Algorithm for Training Text Classifiers”. In Proceedings of the ACM – SIGIR Conference on Research and Development in Information Retrieval, Croft, W. B. and van Rijbergen, C. J. (eds). New York: Springer – Verlag, 1994, pp. 3 – 12

Conformal prediction Permits complementation of predictions made by machine learning algorithms with some measures of reliability Besides the label predicted for a new object, it outputs two additional values – Confidence – Credibility HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions Vovk, V., Gammerman, A. and Shafer, G., Algorithmic Learning in a Random World, New York: Springer Science + Business Media, Inc., 2005

Conformal prediction Used as nonconformity scores the Lagrange multipliers computed during SVM training Extended to a multiclass framework in a one- vs-rest approach HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions

Active learning algorithm Inputs – Initial training set T, calibration set C, pool of candidate samples U – Selection treshold τ, batch size β Train an initial classifier on T While a stopping-criterion is not reached – Apply the current classifier to the pool of samples – Rank the samples in the pool using the uncertainty criterion – Select the top β examples whose certainty level fall under the selection threshold τ – Ask teacher to label the selected samples and add them to the training set – Train a new classifier on the expanded training set HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions

(Un)Certainty criteria Credibility: Confidence: Query-by-transduction: Multicriteria: – Combination 1: – Combination 2: HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions Ho, S. and Wechsler, H., “Query by Transduction”. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 30 (9), 2008, pp – 1571

Setup 1149 samples divided into – Initial training: 5 samples – Pool: 794 samples – Calibration set: 150 samples – Test set: 200 samples Batch size: 25 Selection treshold : 0.4 Used with Qbt, Comb1 and Comb2 HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions

Setup For Each criterion in (Qbt, Comb1, Comb2) – For experiment = 1 : 10 Select test set Run active learning algorithm selecting 700 samples hybridly For NumTrn = 50 : 50 : 700 – Train CP classifier on the first NumTrn samples – Aplly classifier to the test set End for NumTrn – End For experiment End For Each HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions

Results HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions

Results HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions

Conclusions Active learning was applied to the selection of a minimal and good enough training dataset for classification purposes It allows reaching higher success rates and confidence in predictions with fewer data points, compared to the random selection of the training set Combining multiple criteria we can balance the trade-off between success rate and confidence of prediction improvement HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions

ACTIVE LEARNING USING CONFORMAL PREDICTORS: APPLICATION TO IMAGE CLASSIFICATION THANK YOU 7th Workshop on Fusion Data Processing Validation and Analysis HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results HypHyp Conclusions