GAUSSIAN PROCESS REGRESSION WITHIN AN ACTIVE LEARNING SCHEME

Slides:



Advertisements
Similar presentations
Probability and Maximum Likelihood. How are we doing on the pass sequence? This fit is pretty good, but… Hand-labeled horizontal coordinate, t The red.
Advertisements

Design Rule Generation for Interconnect Matching Andrew B. Kahng and Rasit Onur Topaloglu {abk | rtopalog University of California, San Diego.
StatisticalDesign&ModelsValidation. Introduction.
Notes Sample vs distribution “m” vs “µ” and “s” vs “σ” Bias/Variance Bias: Measures how much the learnt model is wrong disregarding noise Variance: Measures.
Secure Unlocking of Mobile Touch Screen Devices by Simple Gestures – You can see it but you can not do it Arjmand Samuel Microsoft Research Muhammad Shahzad.
Department of electrical and computer engineering An Equalization Technique for High Rate OFDM Systems Mehdi Basiri.
Psychology 202b Advanced Psychological Statistics, II February 15, 2011.
Page 0 of 8 Time Series Classification – phoneme recognition in reconstructed phase space Sanjay Patil Intelligent Electronics Systems Human and Systems.
Predictive Automatic Relevance Determination by Expectation Propagation Yuan (Alan) Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani.
NORM BASED APPROACHES FOR AUTOMATIC TUNING OF MODEL BASED PREDICTIVE CONTROL Pastora Vega, Mario Francisco, Eladio Sanz University of Salamanca – Spain.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Implementing a reliable neuro-classifier
1 Channel Estimation for IEEE a OFDM Downlink Transmission Student: 王依翎 Advisor: Dr. David W. Lin Advisor: Dr. David W. Lin 2006/02/23.
1 Ensembles of Nearest Neighbor Forecasts Dragomir Yankov, Eamonn Keogh Dept. of Computer Science & Eng. University of California Riverside Dennis DeCoste.
CHAPTER 4: Parametric Methods. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Parametric Estimation X = {
CHAPTER 4: Parametric Methods. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Parametric Estimation Given.
Distributed and Efficient Classifiers for Wireless Audio-Sensor Networks Baljeet Malhotra Ioanis Nikolaidis Mario A. Nascimento University of Alberta Canada.
Study of Sparse Online Gaussian Process for Regression EE645 Final Project May 2005 Eric Saint Georges.
Remote Sensing Laboratory Dept. of Information Engineering and Computer Science University of Trento Via Sommarive, 14, I Povo, Trento, Italy Remote.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
 Catalogue No: BS-338  Credit Hours: 3  Text Book: Advanced Engineering Mathematics by E.Kreyszig  Reference Books  Probability and Statistics by.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Efficient Direct Density Ratio Estimation for Non-stationarity Adaptation and Outlier Detection Takafumi Kanamori Shohei Hido NIPS 2008.
Minimum Phoneme Error Based Heteroscedastic Linear Discriminant Analysis for Speech Recognition Bing Zhang and Spyros Matsoukas BBN Technologies Present.
A New Subspace Approach for Supervised Hyperspectral Image Classification Jun Li 1,2, José M. Bioucas-Dias 2 and Antonio Plaza 1 1 Hyperspectral Computing.
1 Remote Sensing Laboratory Dept. of Information Engineering and Computer Science University of Trento Via Sommarive, 14, I Povo, Trento, Italy 2.
Least-Mean-Square Training of Cluster-Weighted-Modeling National Taiwan University Department of Computer Science and Information Engineering.
Soft Sensor for Faulty Measurements Detection and Reconstruction in Urban Traffic Department of Adaptive systems, Institute of Information Theory and Automation,
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
Fuzzy Entropy based feature selection for classification of hyperspectral data Mahesh Pal Department of Civil Engineering National Institute of Technology.
Page 1 Ming Ji Department of Computer Science University of Illinois at Urbana-Champaign.
Prognosis of Gear Health Using Gaussian Process Model Department of Adaptive systems, Institute of Information Theory and Automation, May 2011, Prague.
Prognosis of gear health using stochastic dynamical models with online parameter estimation 10th International PhD Workshop on Systems and Control a Young.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Greedy is not Enough: An Efficient Batch Mode Active Learning Algorithm Chen, Yi-wen( 陳憶文 ) Graduate Institute of Computer Science & Information Engineering.
IMPROVING ACTIVE LEARNING METHODS USING SPATIAL INFORMATION IGARSS 2011 Edoardo Pasolli Univ. of Trento, Italy Farid Melgani Univ.
Gaussian Processes Li An Li An
earthobs.nr.no Retraining maximum likelihood classifiers using a low-rank model Arnt-Børre Salberg Norwegian Computing Center Oslo, Norway IGARSS.
Introduction to Pattern Recognition (การรู้จํารูปแบบเบื้องต้น)
GENDER AND AGE RECOGNITION FOR VIDEO ANALYTICS SOLUTION PRESENTED BY: SUBHASH REDDY JOLAPURAM.
Speech Lab, ECE, State University of New York at Binghamton  Classification accuracies of neural network (left) and MXL (right) classifiers with various.
Computational Intelligence: Methods and Applications Lecture 29 Approximation theory, RBF and SFN networks Włodzisław Duch Dept. of Informatics, UMK Google:
Brain-Machine Interface (BMI) System Identification Siddharth Dangi and Suraj Gowda BMIs decode neural activity into control signals for prosthetic limbs.
1 On the Channel Capacity of Wireless Fading Channels C. D. Charalambous and S. Z. Denic School of Information Technology and Engineering, University of.
Learning Photographic Global Tonal Adjustment with a Database of Input / Output Image Pairs.
 Effective Multi-Label Active Learning for Text Classification Bishan yang, Juan-Tao Sun, Tengjiao Wang, Zheng Chen KDD’ 09 Supervisor: Koh Jia-Ling Presenter:
Chapter 15: Classification of Time- Embedded EEG Using Short-Time Principal Component Analysis by Nguyen Duc Thang 5/2009.
RECONSTRUCTION OF MULTI- SPECTRAL IMAGES USING MAP Gaurav.
Presentation : “ Maximum Likelihood Estimation” Presented By : Jesu Kiran Spurgen Date :
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Deep Feedforward Networks
Probability Theory and Parameter Estimation I
Introduction Machine Learning 14/02/2017.
Background on Classification
Transfer Learning in Astronomy: A New Machine Learning Paradigm
CH 5: Multivariate Methods
Alan Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani
Classification Discriminant Analysis
CSCI 5822 Probabilistic Models of Human and Machine Learning
Roberto Battiti, Mauro Brunato
Classification Discriminant Analysis
Unsupervised Learning II: Soft Clustering with Gaussian Mixture Models
Recursively Adapted Radial Basis Function Networks and its Relationship to Resource Allocating Networks and Online Kernel Learning Weifeng Liu, Puskal.
Multivariate Methods Berlin Chen
Physics-guided machine learning for milling stability:
A Data Partitioning Scheme for Spatial Regression
Real-time Uncertainty Output for MBES Systems
Probabilistic Surrogate Models
Presentation transcript:

GAUSSIAN PROCESS REGRESSION WITHIN AN ACTIVE LEARNING SCHEME University of Trento Dept. of Information Engineering and Computer Science Italy GAUSSIAN PROCESS REGRESSION WITHIN AN ACTIVE LEARNING SCHEME Edoardo Pasolli pasolli@disi.unitn.it Farid Melgani melgani@disi.unitn.it July 28, 2011 IGARSS 2011

Introduction Supervised regression approach Pre-processing Feature extraction Regression Image/ Signal Prediction Training sample collection Human expert Training sample quality/quantity Impact on prediction errors

Introduction Active learning approach for classification problems Training (labeled) set Learning (unlabeled) set f2 f2 Training of classifier Active learning method f1 f1 Model of classifier Selected samples from learning (unlabeled) set f2 Insertion in training set Labeling of selected samples Human expert f1 Selected samples after labeling

Objective Propose GP-based active learning strategies for biophysical parameter estimation problems

Gaussian Processes (GPs) Predictive distribution covariance matrix defined by covariance function noise variance

Gaussian Processes (GPs) Example of predicted function : training sample : predicted value : standard deviation of predicted value

Proposed Strategies GP Regression Insertion in Selection training set U: Learning set L: Training set Insertion in training set Selection Human expert L’s: Labeled samples U’s: Selected unlabeled samples Labeling

Proposed Strategies Minimize covariance measure in feature space (Cov) : squared exponential covariance function signal variance length-scale : training sample : covariance function with respect to training sample

Proposed Strategies Minimize covariance measure in feature space (Cov) : training sample : covariance function with respect to training sample : covariance measure with respect to all training samples : selection of samples with minimum values of

Proposed Strategies Maximize variance of predicted value (Var) : training sample : predicted value : standard deviation of predicted value

Proposed Strategies Maximize variance of predicted value (Var) : training sample : variance : selection of samples with maximum values of

Experimental Results Data set description (MERIS) Simulated acquisitions Objective: estimation of chlorophyll concentration in subsurface case I + case II (open and coastal) waters Sensor: MEdium Resolution Imaging Spectrometer (MERIS) # channels: 8 (412-618 nm) Range of chlorophyll concentration: 0.02-54 mg/m3

Experimental Results Data set description (SeaBAM) Real aquisitions Objective: estimation of chlorophyll concentration mostly in subsurface case I (open) waters Sensor: Sea-viewing Wide Field-of-view (SeaWiFS) # channels: 5 (412-555 nm) Range of chlorophyll concentration: 0.02-32.79 mg/m3

Experimental Results Mean Squared Error MERIS SeaBAM

Experimental Results Standard Deviation of Mean Squared Error MERIS SeaBAM

Accuracies on 4000 test samples Experimental Results Detailed results MERIS Accuracies on 4000 test samples Method # training samples MSE σMSE R2 σR2 Full 1000 0.086 - 0.991 Initial 50 1.638 0.869 0.849 0.070 Ran Cov Var 150 0.585 0.378 0.184 0.406 0.105 0.054 0.938 0.961 0.980 0.045 0.010 0.005 300 0.237 0.212 0.095 0.084 0.177 0.975 0.977 0.990 0.008 0.018 0.000

Accuracies on 4000 test samples Experimental Results Detailed results MERIS Accuracies on 4000 test samples Method # training samples MSE σMSE R2 σR2 Full 1000 0.086 - 0.991 Initial 50 1.638 0.869 0.849 0.070 Ran Cov Var 150 0.585 0.378 0.184 0.406 0.105 0.054 0.938 0.961 0.980 0.045 0.010 0.005 300 0.237 0.212 0.095 0.084 0.177 0.975 0.977 0.990 0.008 0.018 0.000

Accuracies on 459 test samples Experimental Results Detailed results SeaBAM Accuracies on 459 test samples Method # training samples MSE σMSE R2 σR2 Full 460 1.536 - 0.806 Initial 60 5.221 2.968 0.526 0.215 Ran Cov Var 160 2.972 2.210 1.818 1.038 0.074 0.029 0.682 0.745 0.784 0.069 0.007 0.003 310 2.062 1.601 1.573 0.687 0.010 0.753 0.800 0.803 0.066 0.001 0.000

Accuracies on 459 test samples Experimental Results Detailed results SeaBAM Accuracies on 459 test samples Method # training samples MSE σMSE R2 σR2 Full 460 1.536 - 0.806 Initial 60 5.221 2.968 0.526 0.215 Ran Cov Var 160 2.972 2.210 1.818 1.038 0.074 0.029 0.682 0.745 0.784 0.069 0.007 0.003 310 2.062 1.601 1.573 0.687 0.010 0.753 0.800 0.803 0.066 0.001 0.000

Conclusions In this work, GP-based active learning strategies for regression problems are proposed Encouraging performances in terms of convergence speed stability Future developments extension to other regression approaches

GAUSSIAN PROCESS REGRESSION WITHIN AN ACTIVE LEARNING SCHEME University of Trento Dept. of Information Engineering and Computer Science Italy GAUSSIAN PROCESS REGRESSION WITHIN AN ACTIVE LEARNING SCHEME Edoardo Pasolli pasolli@disi.unitn.it Farid Melgani melgani@disi.unitn.it July 28, 2011 IGARSS 2011