Part 2: Change detection for SAR Imagery (based on Chapter 18 of the Book. Change-detection methods for location of mines in SAR imagery, by Dr. Nasser.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Principles of Density Estimation
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
An Overview of Machine Learning
Real-Time Human Pose Recognition in Parts from Single Depth Images Presented by: Mohammad A. Gowayyed.
Paper presentation for CSI5388 PENGCHENG XI Mar. 23, 2005
Multiple Criteria for Evaluating Land Cover Classification Algorithms Summary of a paper by R.S. DeFries and Jonathan Cheung-Wai Chan April, 2000 Remote.
Chapter 5: Linear Discriminant Functions
Announcements  Project proposal is due on 03/11  Three seminars this Friday (EB 3105) Dealing with Indefinite Representations in Pattern Recognition.
Statistical Decision Theory, Bayes Classifier
Ensemble Learning: An Introduction
Three kinds of learning
Presented by Zeehasham Rasheed
Chapter 4 (part 2): Non-Parametric Classification
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Statistical Learning: Pattern Classification, Prediction, and Control Peter Bartlett August 2002, UC Berkeley CIS.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Ensemble Learning (2), Tree and Forest
Data Selection In Ad-Hoc Wireless Sensor Networks Olawoye Oyeyele 11/24/2003.
Image Classification and its Applications
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Outline Separating Hyperplanes – Separable Case
earthobs.nr.no Land cover classification of cloud- and snow-contaminated multi-temporal high-resolution satellite images Arnt-Børre Salberg and.
Data mining and machine learning A brief introduction.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
A New Subspace Approach for Supervised Hyperspectral Image Classification Jun Li 1,2, José M. Bioucas-Dias 2 and Antonio Plaza 1 1 Hyperspectral Computing.
Ajay Kumar, Member, IEEE, and David Zhang, Senior Member, IEEE.
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
1 SUPPORT VECTOR MACHINES İsmail GÜNEŞ. 2 What is SVM? A new generation learning system. A new generation learning system. Based on recent advances in.
Image Classification 영상분류
1 Part II: Practical Implementations.. 2 Modeling the Classes Stochastic Discrimination.
EECS 274 Computer Vision Segmentation by Clustering II.
ECE738 Advanced Image Processing Face Detection IEEE Trans. PAMI, July 1997.
Today Ensemble Methods. Recap of the course. Classifier Fusion
1 E. Fatemizadeh Statistical Pattern Recognition.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Pattern Recognition April 19, 2007 Suggested Reading: Horn Chapter 14.
1 Pattern Recognition Pattern recognition is: 1. A research area in which patterns in data are found, recognized, discovered, …whatever. 2. A catchall.
Chapter 4: Pattern Recognition. Classification is a process that assigns a label to an object according to some representation of the object’s properties.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
MACHINE LEARNING 10 Decision Trees. Motivation  Parametric Estimation  Assume model for class probability or regression  Estimate parameters from all.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Chapter 13 (Prototype Methods and Nearest-Neighbors )
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Mixture Densities Maximum Likelihood Estimates.
Cluster Analysis What is Cluster Analysis? Types of Data in Cluster Analysis A Categorization of Major Clustering Methods Partitioning Methods.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
Nonparametric Density Estimation – k-nearest neighbor (kNN) 02/20/17
LECTURE 11: Advanced Discriminant Analysis
Ch8: Nonparametric Methods
Machine Learning Basics
Outline Parameter estimation – continued Non-parametric methods.
REMOTE SENSING Multispectral Image Classification
Computer Vision Chapter 4
LECTURE 21: CLUSTERING Objectives: Mixture Densities Maximum Likelihood Estimates Application to Gaussian Mixture Models k-Means Clustering Fuzzy k-Means.
Model generalization Brief summary of methods
Hairong Qi, Gonzalez Family Professor
Machine Learning with Clinical Data
Hairong Qi, Gonzalez Family Professor
ECE – Pattern Recognition Lecture 10 – Nonparametric Density Estimation – k-nearest-neighbor (kNN) Hairong Qi, Gonzalez Family Professor Electrical.
Hairong Qi, Gonzalez Family Professor
Presentation transcript:

Part 2: Change detection for SAR Imagery (based on Chapter 18 of the Book. Change-detection methods for location of mines in SAR imagery, by Dr. Nasser Nasrabadi, et al.) * Let x and y be vectors from reference and test images respectively Distance (Euclidean, Mahalanobis) based change detection. Image ratioing Subspace projection-based change detection here P x is a subspace projection operator. Wiener prediction-based change detection. here y w is the Wiener filtered image vector. The following 15 slides are kindly provided by Dr. Nasser Nasrabadi.

Additional remarks Results indicate that in general the more complex algorithms exhibited superior performance as compared to the simple methods; the exception seems to be the ratio method. The ratio method, although a simple implementation, has performance which is competitive with the more complex change detection methods implemented. Results indicate that taking into account the local spatial and statistical properties of the image improves the change detection performance. The ROC curves also show that addition of the term may serve to mitigate false alarms that arise from speckle. As expected the simple difference method exhibits the worst performance, having a high occurrence of false alarms. The authors also examined Kernel versions of selected change detection methods. The results are based on very limited data.

Advanced approaches to change detection in SAR images (based on chapter 17 of the Book, Unsupervised change detection in muti-temporal SAR images, by Dr. Bruzzone, et al.) Nice literature survey Deals with images of the same scenes at two different time instants. The proposed approach includes: Multi-resolution decomposition of log-ratio image. Adaptive scale identification Scale-driven fusion.

Part 3a : The Classification Problems Based on Chapters 2,3,16,17,22,25,,26,and 27 of the Book Pixel classification and segmentation into regions are typical image classification problems. Event classification from signals is another popular classification problem. * Statistical pattern classification is still very important. The use of ensemble classifiers is popular. * Neural networks and support vector machines emerge as more practical approaches to the classification problems.

Statistical Classifiers Advantages of statistical classifiers: * Theory and techniques are now well developed. * The parametric classification rules include the (optimum) Bayes, maximum likelihood decision rules and discriminant analysis. The nonparametric (or distribution free) method of classification includes nearest neighbor decision rule and Parzen’s density estimate. * Several hundred papers on NNDR have been published. One variation of NNDR is called the voting k near centroid neighborhood rule (k-NCN). An improvement of 1 to 1.5% over k-NNDR can be achieved. Still the original k-NNDR performs well with proper reference data sets. (S. Grqabowski, et al., “Nearest neighbor decision rule for pixel classification in remote sensing, in Frontiers of Remote Sensing Information Processing. Feature extraction and selection algorithms have been well studied. Contextual information can be incorporated in the decision rule. Remark: To seek for improved classifier is always a challenge.

Restrictions of Statistical Classifiers Accuracy of the statistical information is always of concern especially when the number of training samples is small. One solution to “enhance” the statistics is to use unsupervised samples. (work by Prof. Landgrebe). The “Hughes phenomenon”: For a fixed training sample size, the classification performance first improves with the increase in number of features until it reaches a peak beyond which further increase in feature number will degrade the performance. Difficulty of using higher than second order statistics. Most algorithms assume underlying Gaussian statistics. Spatial dependence information cannot be fully captured by statistics. Syntactic, semantic, and structural pattern recognition approaches have their useful role though statistical classifiers are most popular.

Random Forest (RF) Classification of Remote Sensing Data (based on Chapter 15 of the Book) A random forest (RF) classifier is a classifier comprising of a collection of classification trees, each trained on a randomly chosen subset of the input data where final classification is based on a majority vote by the trees in the forest. Each tree is trained using a random subset of the training samples. During the growing process of a tree the best split on each node in the tree is found by searching through m randomly selected features. For a data set with M features, m is selected by the user and kept much smaller then M. Every tree is grown to its fullest to diversify the trees so there is no pruning. RF was investigated for hyperspectral and muti-source remote sensing data.

RF classifiers continued-1 Random forests have been shown to be comparable to boosting in terms of accuracies, but are computationally much less intensive than boosting. Boosting is an ensemble classification method that train several classifiers and combine their results through a voting process. Two random forest (RF) approaches are explored; the RF-CART (Classification and Regression Tree) and the RF-BHC (Binary Hierarchical Classifier). The RF-CART approach is based on CART-like trees where trees are grown to minimize an impurity measure. The RF-BHC approach depends on class separability measures and the Fisher projection, which maximizes the Fisher discriminant where each tree is a class hierarchy, and the number of leaves is the same as the number of classes.

RF classifiers continued-2 The different RF approaches are compared in experiments in classification of a multisource remote sensing and geographic data from the Anderson River in Canada and urban area from Pavia, Italy using hyperspectral ROSIS data. RF is attractive for both types of data. Both approaches perform well. For ROSIS data,RF-BHC performs better than BHC and RF-CART in test data RF approach is fast in training and classification and is distribution free. * The problem of curse of dimensionality is addressed by the selection of low m.

ROSIS data: reference image (left), gray scale image (right)

The BHC tree used for classification of ROSIS data.