Active learning The learning algorithm must have some control over the data from which it learns It must be able to query an oracle, requesting for labels.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Active Learning with Feedback on Both Features and Instances H. Raghavan, O. Madani and R. Jones Journal of Machine Learning Research 7 (2006) Presented.
Sequential Minimal Optimization Advanced Machine Learning Course 2012 Fall Semester Tsinghua University.
DECISION TREES. Decision trees  One possible representation for hypotheses.
+ Multi-label Classification using Adaptive Neighborhoods Tanwistha Saha, Huzefa Rangwala and Carlotta Domeniconi Department of Computer Science George.
Data Mining Feature Selection. Data reduction: Obtain a reduced representation of the data set that is much smaller in volume but yet produces the same.
Supporting top-k join queries in relational databases Ihab F. Ilyas, Walid G. Aref, Ahmed K. Elmagarmid Presented by Rebecca M. Atchley Thursday, April.
Support Vector Machines
Machine learning continued Image source:
Learning Globally-Consistent Local Distance Functions for Shape-Based Image Retrieval and Classification Computer Vision, ICCV IEEE 11th International.
Proactive Learning: Cost- Sensitive Active Learning with Multiple Imperfect Oracles Pinar Donmez and Jaime Carbonell Pinar Donmez and Jaime Carbonell Language.
Face Recognition & Biometric Systems Support Vector Machines (part 2)
Maria-Florina Balcan Modern Topics in Learning Theory Maria-Florina Balcan 04/19/2006.
Supervised learning Given training examples of inputs and corresponding outputs, produce the “correct” outputs for new inputs Two main scenarios: –Classification:
Fuzzy Support Vector Machines (FSVMs) Weijia Wang, Huanren Zhang, Vijendra Purohit, Aditi Gupta.
1 Algorithms for Large Data Sets Ziv Bar-Yossef Lecture 8 May 4, 2005
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
Reduced Support Vector Machine
Active Learning with Support Vector Machines
Unsupervised Training and Clustering Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
Bioinformatics Challenge  Learning in very high dimensions with very few samples  Acute leukemia dataset: 7129 # of gene vs. 72 samples  Colon cancer.
DUAL STRATEGY ACTIVE LEARNING presenter: Pinar Donmez 1 Joint work with Jaime G. Carbonell 1 & Paul N. Bennett 2 1 Language Technologies Institute, Carnegie.
Active Learning Strategies for Compound Screening Megon Walker 1 and Simon Kasif 1,2 1 Bioinformatics Program, Boston University 2 Department of Biomedical.
R-CNN By Zhang Liliang.
Introduction to Machine Learning Approach Lecture 5.
Ensemble Learning (2), Tree and Forest
Computational Analysis of USA Swimming Data Junfu Xu School of Computer Engineering and Science, Shanghai University.
1 Efficiently Learning the Accuracy of Labeling Sources for Selective Sampling by Pinar Donmez, Jaime Carbonell, Jeff Schneider School of Computer Science,
MediaEval Workshop 2011 Pisa, Italy 1-2 September 2011.
SVM by Sequential Minimal Optimization (SMO)
Support Vector Machines Mei-Chen Yeh 04/20/2010. The Classification Problem Label instances, usually represented by feature vectors, into one of the predefined.
Active Learning on Spatial Data Christine Körner Fraunhofer AIS, Uni Bonn.
ACTIVE LEARNING USING CONFORMAL PREDICTORS: APPLICATION TO IMAGE CLASSIFICATION HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results.
Classification and Ranking Approaches to Discriminative Language Modeling for ASR Erinç Dikici, Murat Semerci, Murat Saraçlar, Ethem Alpaydın 報告者:郝柏翰 2013/01/28.
Semi-supervised Learning on Partially Labeled Imbalanced Data May 16, 2010 Jianjun Xie and Tao Xiong.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Binxing Jiao et. al (SIGIR ’10) Presenter : Lin, Yi-Jhen Advisor: Dr. Koh. Jia-ling Date: 2011/4/25 VISUAL SUMMARIZATION OF WEB PAGES.
Greedy is not Enough: An Efficient Batch Mode Active Learning Algorithm Chen, Yi-wen( 陳憶文 ) Graduate Institute of Computer Science & Information Engineering.
Wikipedia as Sense Inventory to Improve Diversity in Web Search Results Celina SantamariaJulio GonzaloJavier Artiles nlp.uned.es UNED,c/Juan del Rosal,
1 Opinion Retrieval from Blogs Wei Zhang, Clement Yu, and Weiyi Meng (2007 CIKM)
Evolutionary Algorithms for Finding Optimal Gene Sets in Micro array Prediction. J. M. Deutsch Presented by: Shruti Sharma.
Non-Bayes classifiers. Linear discriminants, neural networks.
Confidence Based Autonomy: Policy Learning by Demonstration Manuela M. Veloso Thanks to Sonia Chernova Computer Science Department Carnegie Mellon University.
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
Active learning Haidong Shi, Nanyi Zeng Nov,12,2008.
Effective Automatic Image Annotation Via A Coherent Language Model and Active Learning Rong Jin, Joyce Y. Chai Michigan State University Luo Si Carnegie.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Assignments CS fall Assignment 1 due Generate the in silico data set of 2sin(1.5x)+ N (0,1) with 100 random values of x between.
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
NTU & MSRA Ming-Feng Tsai
 Effective Multi-Label Active Learning for Text Classification Bishan yang, Juan-Tao Sun, Tengjiao Wang, Zheng Chen KDD’ 09 Supervisor: Koh Jia-Ling Presenter:
Gist 2.3 John H. Phan MIBLab Summer Workshop June 28th, 2006.
Incremental Reduced Support Vector Machines Yuh-Jye Lee, Hung-Yi Lo and Su-Yun Huang National Taiwan University of Science and Technology and Institute.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Data Mining Practical Machine Learning Tools and Techniques
CS 9633 Machine Learning Support Vector Machines
Cluster Analysis II 10/03/2012.
Adversarial Learning for Neural Dialogue Generation
Classification with Perceptrons Reading:
Machine Learning Week 1.
Ying shen Sse, tongji university Sep. 2016
Tagging Review Comments Rationale #10 Week 13
Supervised vs. unsupervised Learning
Learning a Policy for Opportunistic Active Learning
Ensemble learning.
Using Clustering to Make Prediction Intervals For Neural Networks
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
Presentation transcript:

Active learning The learning algorithm must have some control over the data from which it learns It must be able to query an oracle, requesting for labels of data samples that seem to be most informative for the learning process Proper selection of samples implies better performances with fewer data

Scenarios Learning with membership queries Stream-based sampling Pool-based sampling

Strategies Uncertainty sampling Query-by-committee Density-weighted…

Conformal prediction Permits complementation of predictions made by machine learning algorithms with some measures of reliability The label predicted for a new object must make it similar to the old objects The degree of similarity is used to estimate the confidence in the prediction

Conformal prediction algorithm Inputs: Training sample and a test sample Consider all possible values for the label ; Compute nonconformity scores and p-values for each possible classification; Predict the label corresponding to the largest p-value calculated; Output one minus the second largest p-value as the confidence for the prediction; Output the largest p-value calculated as the credibility of the prediction.

Nonconformity scores and p-values Used as nonconformity scores the Lagrange multipliers computed during SVM training Extended to a multiclass framework in a one-vs-rest approach P-values:

Active learning algorithm Inputs Initial training set T, calibration set C, pool of candidate samples U Selection treshold τ, batch size β Train an initial classifier on T While a stopping-criterion is not reached Apply the current classifier to the pool of samples Rank the samples in the pool using the uncertainty criterion Select the top β examples whose certainty level fall under the selection threshold τ Ask teacher to label the selected examples and add them to the training set Train a new classifier on the expanded training set

Stopping criteria Pre-specified size for the training set Exhaustion of the pool of candidate samples Early-stop Implemented using the calibration set Active selection stops if no improvements can be obtained when applying newly trained classifiers to the calibration set