1 Keng-Hao Liu and Chein-I Chang Remote Sensing Signal and Image Processing Laboratory (RSSIPL) Department of Computer Science and Electrical Engineering.

Slides:



Advertisements
Similar presentations
Brief introduction on Logistic Regression
Advertisements

Object Specific Compressed Sensing by minimizing a weighted L2-norm A. Mahalanobis.
Huffman code and ID3 Prof. Sin-Min Lee Department of Computer Science.
A Graphical Operator Framework for Signature Detection in Hyperspectral Imagery David Messinger, Ph.D. Digital Imaging and Remote Sensing Laboratory Chester.
Experimental Design, Response Surface Analysis, and Optimization
Major Operations of Digital Image Processing (DIP) Image Quality Assessment Radiometric Correction Geometric Correction Image Classification Introduction.
Hyperspectral Imagery
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Chapter 2: Pattern Recognition
Machine Learning CMPT 726 Simon Fraser University
Independent Component Analysis (ICA) and Factor Analysis (FA)
Control Charts for Variables
Lecture 14: Classification Thursday 18 February 2010 Reading: Ch – 7.19 Last lecture: Spectral Mixture Analysis.
Lecture 11 Multivariate Regression A Case Study. Other topics: Multicollinearity  Assuming that all the regression assumptions hold how good are our.
Classification and Prediction by Yen-Hsien Lee Department of Information Management College of Management National Sun Yat-Sen University March 4, 2003.
Lecture 14: Classification Thursday 19 February Reading: “Estimating Sub-pixel Surface Roughness Using Remotely Sensed Stereoscopic Data” pdf preprint.
Digital Imaging and Remote Sensing Laboratory Real-World Stepwise Spectral Unmixing Daniel Newland Dr. John Schott Digital Imaging and Remote Sensing Laboratory.
Control Charts for Attributes
Determining the Size of
Ensemble Learning (2), Tree and Forest
Software Research Image Compression Mohamed N. Ahmed, Ph.D.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Remote Sensing Hyperspectral Remote Sensing. 1. Hyperspectral Remote Sensing ► Collects image data in many narrow contiguous spectral bands through the.
Noise-Robust Spatial Preprocessing Prior to Endmember Extraction from Hyperspectral Data Gabriel Martín, Maciel Zortea and Antonio Plaza Hyperspectral.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
Statistics and Quantitative Analysis Chemistry 321, Summer 2014.
National Aerospace University “Kharkov Aviation Institute” SPIE Remote Sensing Performance prediction for 3D filtering of multichannel images Oleksii.
Relationships between Land Cover and Spatial Statistical Compression in High-Resolution Imagery James A. Shine 1 and Daniel B. Carr 2 34 th Symposium on.
Fuzzy Entropy based feature selection for classification of hyperspectral data Mahesh Pal Department of Civil Engineering National Institute of Technology.
Remote Sensing Supervised Image Classification. Supervised Image Classification ► An image classification procedure that requires interaction with the.
Jonatan Gefen 28/11/2012. Outline Introduction to classification Whole Pixel Subpixel Classification Linear Unmixing Matched Filtering (partial unmixing)
1 Particle Swarm Optimization-based Dimensionality Reduction for Hyperspectral Image Classification He Yang, Jenny Q. Du Department of Electrical and Computer.
EMIS 7300 SYSTEMS ANALYSIS METHODS FALL 2005 Dr. John Lipp Copyright © Dr. John Lipp.
Computer Vision – Compression(1) Hanyang University Jong-Il Park.
Digital Image Processing Definition: Computer-based manipulation and interpretation of digital images.
IGARSS 2011, Vancouver, Canada HYPERSPECTRAL UNMIXING USING A NOVEL CONVERSION MODEL Fereidoun A. Mianji, Member, IEEE, Shuang Zhou, Member, IEEE, Ye Zhang,
Advances in digital image compression techniques Guojun Lu, Computer Communications, Vol. 16, No. 4, Apr, 1993, pp
Multiple Antennas Have a Big Multi- User Advantage in Wireless Communications Bertrand Hochwald (Bell Labs)
14 ARM Science Team Meeting, Albuquerque, NM, March 21-26, 2004 Canada Centre for Remote Sensing - Centre canadien de télédétection Geomatics Canada Natural.
Validation of MODIS Snow Mapping Algorithm Jiancheng Shi Institute for Computational Earth System Science University of California, Santa Barbara.
Lecture 4: Lossless Compression(1) Hongli Luo Fall 2011.
Remote Sensing Unsupervised Image Classification.
Review of Spectral Unmixing for Hyperspectral Imagery Lidan Miao Sept. 29, 2005.
Digital Imaging and Remote Sensing Laboratory Maximum Noise Fraction Transform 1 Generate noise covariance matrix Use this to decorrelate (using principle.
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
Various Change Detection Analysis Techniques. Broadly Divided in Two Approaches ….. 1.Post Classification Approach. 2.Pre Classification Approach.
The Design of Statistical Specifications for a Test Mark D. Reckase Michigan State University.
Sub pixelclassification
Comparative Analysis of Spectral Unmixing Algorithms Lidan Miao Nov. 10, 2005.
Principal Component Analysis
Huffman Coding (2 nd Method). Huffman coding (2 nd Method)  The Huffman code is a source code. Here word length of the code word approaches the fundamental.
The inference and accuracy We learned how to estimate the probability that the percentage of some subjects in the sample would be in a given interval by.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Model Comparison. Assessing alternative models We don’t ask “Is the model right or wrong?” We ask “Do the data support a model more than a competing model?”
Chapter 7. Classification and Prediction
Decision Trees an introduction.
ECO 173 Chapter 10: Introduction to Estimation Lecture 5a
Orthogonal Subspace Projection - Matched Filter
Classification of unlabeled data:
Incorporating Ancillary Data for Classification
ECO 173 Chapter 10: Introduction to Estimation Lecture 5a
University College London (UCL), UK
Hyperspectral Image preprocessing
REMOTE SENSING Multispectral Image Classification
REMOTE SENSING Multispectral Image Classification
Image Information Extraction
Igor Appel Alexander Kokhanovsky
Principal Component Analysis
Parametric Methods Berlin Chen, 2005 References:
Principal Component Analysis (PCA)
Presentation transcript:

1 Keng-Hao Liu and Chein-I Chang Remote Sensing Signal and Image Processing Laboratory (RSSIPL) Department of Computer Science and Electrical Engineering University of Maryland, Baltimore County (UMBC) Baltimore, MD Dynamic Band Selection For Hyperspectral Imagery

2 Motivation Band Selection (BS) is one of commonly used approaches that take advantage of high inter- band correlation to remove band redundancy in order to achieve a wide range of applications. Band Selection (BS) is one of commonly used approaches that take advantage of high inter- band correlation to remove band redundancy in order to achieve a wide range of applications. However, there are several crucial issues arising in implementation of BS. One of these issues is how to estimate the number of bands, p, required to be selected? However, there are several crucial issues arising in implementation of BS. One of these issues is how to estimate the number of bands, p, required to be selected? How to find p? How to find p?

3 Outline Dynamic Band Selection (DBS) Dynamic Band Selection (DBS) - Virtual Dimensionality (VD) - Band Dimensionality Allocation (BDA) - Progressive Band Dimensionality Process (PBDP) - Criteria for Band Prioritization (BP) Experiments Experiments - HYDICE - AVIRIS (Purdue) Data Conclusions Conclusions

4 Issues of Conventional BS Require knowing the number of bands required for BS, p, a priori The value of p is fixed at a constant and cannot be adaptive. Need an exhaustive search required to find an optimal set of p bands among all possible combinations out of the total number of bands.

5 Progressive Band Dimensionality Process (PBDP) Using a criterion prioritizes all L spectral bands and removes highly correlated bands then selects bands progressively in a forward or backward manner depending upon how to retain band information in increasing or decreasing order. Using a criterion prioritizes all L spectral bands and removes highly correlated bands then selects bands progressively in a forward or backward manner depending upon how to retain band information in increasing or decreasing order.  Forward Progressive Band Dimensionality Process (FPBDP)  Backward Progressive Band Dimensionality Process (BPBDP) The PBDP process is continued on until it reaches a specific number of bands, p. So the p is considered as variable instead of fixed value. The PBDP process is continued on until it reaches a specific number of bands, p. So the p is considered as variable instead of fixed value.

6 Band Prioritization (BP) Criteria for PBDP Band Prioritization Criteria for PBDP Band Prioritization Criteria for PBDP Second order statistic-based BP criteria - Variance - Signal-to-Noise Ratio (SNR or MNF) High order statistic-based BP criteria - Skewness - Kurtosis Infinite order Statistics BP criteria - Entropy - Information Divergence (ID) ) - Neg-entropy (combination of 3rd and 4th order)

7 Virtual Dimensionality (VD) Use VD [Chang 2003] to determine the number of components required for Hyperspectral images. Use VD [Chang 2003] to determine the number of components required for Hyperspectral images. We assume one spectrally distinct signature can be accommodated by one band. So the number of bands required to be selected must be equal or greater than the VD. We assume one spectrally distinct signature can be accommodated by one band. So the number of bands required to be selected must be equal or greater than the VD.

Band Dimensionality Allocation (BDA) Concept is derived from information theory where a source S is emitted by a set of source alphabets that are used to represent the source with a given probability distribution where p j is the probability of the occurrence of the source alphabet a j Concept is derived from information theory where a source S is emitted by a set of source alphabets that are used to represent the source with a given probability distribution where p j is the probability of the occurrence of the source alphabet a j Similarly, assumes m j is material substance signature to be analyzed, then the n j denotes the number of components (bands) required to represent m j. In other word, n j is actually determined by how difficult the m j is discriminated in terms of spectral similarity. Similarly, assumes m j is material substance signature to be analyzed, then the n j denotes the number of components (bands) required to represent m j. In other word, n j is actually determined by how difficult the m j is discriminated in terms of spectral similarity. In conventional BS, it assumes n j =p for all signatures. In this cases all substance are assumed to have equal difficulty to be discriminated by spectral similarity. Generally it is not true in hyperspectral data. In conventional BS, it assumes n j =p for all signatures. In this cases all substance are assumed to have equal difficulty to be discriminated by spectral similarity. Generally it is not true in hyperspectral data. 8

9 Band Dimensionality Allocation (BDA) for signatures Select a spectral similarity measure and a reference signature s. Calculate spectral similarity values for each signature and normalize them to a probability vector. Find self-information. Find the smallest integer q j larger than these self- information. Define dimensionality allocation then assigned it to jth signature, m j. BDA Procedures: Determines the number of signatures to be used for data analysis Additional number required for m j to distinguish itself from other signatures.

10 Band Dimensionality Allocation (BDA) for signatures Three techniques used to find BDA Three techniques used to find BDA   Shannon coding   Huffman coding   Hamming coding Candidates that can be used for spectral similarity measure: Candidates that can be used for spectral similarity measure: Spectral angle mapper (SAM) Spectral information divergence (SID) Candidates that can be used as reference signature (s): Candidates that can be used as reference signature (s): Data sample mean Signature mean or any signature

11 Dynamic Band Selection (DBS) Custom design a criterion for Band Prioritization (BP) Custom design a criterion for Band Prioritization (BP) Implement PBDP Implement PBDP Apply Band De-correlation (BD) Apply Band De-correlation (BD) Band Dimensionality Allocation (BDA) Band Dimensionality Allocation (BDA) DBS steps:

12 Hyperspectral Images Used for Experiments (1) HYDICE Data: 64x bands hyperspectral image with spatial resolution is 20m. Ground truth (desired signatures) Image scene p 11, p 12, p 13 p 211, p 22, p 23 p 221 p 311, p 312, p 32, p 33 p 411, p 412, p 42, p 43 p 511, p 52, p 53 p 521 interferer grass tree road undesired signatures Spectral of five panels Classifier: FCLS Band De-correlation (BD) is applied after BP with σ= 0.1

13 HYDICE Data Experiments signature m j n VD πjπj qjqj n j (BDA) Shannon coding m 1 =p 1 (panels in row1) 9 SID SAM m 2 =p 2 (panels in row2) 9 SID SAM m 3 =p 3 (panels in row3) 9 SID SAM m 4 =p 4 (panels in row4) 9 SID SAM m 5 =p 5 (panels in row5) 9 SID SAM m 6 (grass) 9 SID SAM m 7 (road) 9 SID SAM m 8 (tree) 9 SID SAM m 9 (interferer) 9 SID SAM Shannon BDA results of HYDICE Data

14 HYDICE Data Experiments Unmixed abundance fractions of 19 panel pixels by FCLS n VD Shannon codingHuffman codingHamming coding2n VD totalOptimal p = Number of selected bands to 70 p 11 Variance (65)0.78(32) Skewness (70)0.92(36) Entropy (68)1(24) p 12 Variance (65)0.68(11) Skewness (70)0.68(38) Entropy (68)0.93(10) p 13 Variance000000(65)0(10) Skewness (70)0.23(34) Entropy (68)0.83(12) p = Number of selected bands p 211 Variance (65) Skewness (70)0.99(36) Entropy (68)1.2(9) p 221 Variance (65) Skewness (70)1(37) Entropy (68)1(38) p 22 Variance (65)0.86(14) Skewness (70)0.87(9) Entropy (68)0.78(67) p 23 Variance (65)0.49(14) Skewness (70) Entropy (68) VDBDA

HYDICE Data Experiments ROC performance of 5 row panels using FCLS Panels in row 1Panels in row 2Panels in row 3 Panels in row 4 Panels in row 5 BDA VD 15 Y axis: Area under curve of ROC (P D versus P F ) X axis: Number of selected bands, p

16 HYDICE Data Experiments Average ROC performance of 5 row panels Average performance of 5 row panels BDA range VD 2VD

17 Some notes for HYDICE Data Experiments Classifying subpixels panels requires more bands than pure pixels. High order statistics BPC generally requires a smaller number of bands than 2nd order statistic BPC. The skewness seems to work the best for HYDICE data. To unmix panels, p=n VD =9 seems to be insufficient. But they achieve considerable performance within p=2n VD =18. It implies that the BDA provides a better way to predict cut-off band than VD.

18 Hyperspectral Images Used for Experiments (2) Class map AVIRIS (Purdue) Data: 145x bands hyperspectral image. Image scene class1class2class3class4class5class6class7class8class9class10 class11class12class13class14class15class16 17 Classes maps Data samples are heavily-mixed Classifier: MLC Band De-correlation (BD) is applied after BP with σ= 0.1

19 Purdue Data Experiments BDA results of Purdue Data signature m j n VD πjπj q j j n j (BDA) Shannon coding m 1 (class 1)29SID m 2 (class 2)29SID m 3 (class 3)29SID m 4 (class 4)29SID m 5 (class 5)29SID m 6 (class 6)29SID m 7 (class 7)29SID m 8 (class 8)29SID m 9 (class 9)29SID m 10 (class 10)29SID m 11 (class 11)29SID m 12 (class 12)29SID m 13 (class 13)29SID m 14 (class 14)29SID m 15 (class 15)29SID m 16 (class 16)29SID m 17 (BKG)29SID

20 Purdue Data Experiment MLC classification results of 16 classes class1class2class3class4class5 class6class7class8class9class10 BDA VD Y axis: MLC classification rate in percent% X axis: Number of selected bands, p Classes 1 to 10

21 Purdue Data Experiment Average performance of 16 classes MLC classification results class11class12class13class14class15 class16 VD 2VD BDA range Classes 11 to 16

22 Some Notes for Purdue Data Experiments Second order statistic BPC generally perform better than High order statistic BPC due to - the land covers of this particular scene are large - the data samples are heavily mixed because of low spatial resolution and their contributions to statistics are mainly 2nd order statistics In most of classes using fewer dimensions for MLC can perform as well the using all bands. For instance, classes 7, 9, 13, and 16 do not require more bands to produce the best results. In most of classes using fewer dimensions for MLC can perform as well the using all bands. For instance, classes 7, 9, 13, and 16 do not require more bands to produce the best results. Only 5 classes, 2, 3, 4, 8, and 15 which required almost full dimensions to produce the best MLC results. Only 5 classes, 2, 3, 4, 8, and 15 which required almost full dimensions to produce the best MLC results. DBS provide some guidelines in selecting appropriate p for MLC to perform reasonably. DBS provide some guidelines in selecting appropriate p for MLC to perform reasonably.

23 Summary of DBS The DBS is achieved by implementing the PBDP conjunction with BDA. The DBS is achieved by implementing the PBDP conjunction with BDA. DBS provides a guideline to decide how many bands is needed for each different signature. DBS provides a guideline to decide how many bands is needed for each different signature. The selection of BP criteria has huge influence on the unmixing/classification results. Different applications may requires different BP criteria to produce the best performance. The selection of BP criteria has huge influence on the unmixing/classification results. Different applications may requires different BP criteria to produce the best performance. VD is indeed a good estimate. VD is indeed a good estimate.

24 Progressive Band Dimensionality Process (PBDP) provides a way to estimate p adaptively by increasing bands in a forward manner and decreasing bands in a backward manner. Progressive Band Dimensionality Process (PBDP) provides a way to estimate p adaptively by increasing bands in a forward manner and decreasing bands in a backward manner. Since various material substance signatures require different values of the p for data processing, the Band Dimensionality Allocation (BDA) is further developed to determine different numbers of spectral bands required by individual signatures. Since various material substance signatures require different values of the p for data processing, the Band Dimensionality Allocation (BDA) is further developed to determine different numbers of spectral bands required by individual signatures. Conclusions

25 Thank You !