Environmental Remote Sensing Lecture 5: Image Classification " Purpose: – categorising data – data abstraction / simplification – data interpretation –

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Major Operations of Digital Image Processing (DIP) Image Quality Assessment Radiometric Correction Geometric Correction Image Classification Introduction.
Supervised Learning Recap
Bayesian Decision Theory
ASTER image – one of the fastest changing places in the U.S. Where??
Multiple Criteria for Evaluating Land Cover Classification Algorithms Summary of a paper by R.S. DeFries and Jonathan Cheung-Wai Chan April, 2000 Remote.
Chapter 2: Pattern Recognition
An Overview of RS Image Clustering and Classification by Miles Logsdon with thanks to Robin Weeks Frank Westerlund.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Image Classification.
Lecture 4 Unsupervised Learning Clustering & Dimensionality Reduction
Basics: Notation: Sum:. PARAMETERS MEAN: Sample Variance: Standard Deviation: * the statistical average * the central tendency * the spread of the values.
Lecture 14: Classification Thursday 18 February 2010 Reading: Ch – 7.19 Last lecture: Spectral Mixture Analysis.
Image Classification To automatically categorize all pixels in an image into land cover classes or themes.
Lecture 14: Classification Thursday 19 February Reading: “Estimating Sub-pixel Surface Roughness Using Remotely Sensed Stereoscopic Data” pdf preprint.
Image Classification: Supervised Methods
Image Classification: Redux
METU Informatics Institute Min 720 Pattern Classification with Bio-Medical Applications PART 2: Statistical Pattern Classification: Optimal Classification.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Image Classification
Data Mining Techniques
Classification Advantages of Visual Interpretation: 1. Human brain is the best data processor for information extraction, thus can do much more complex.
Image Classification and its Applications
1 Linear Methods for Classification Lecture Notes for CMPUT 466/551 Nilanjan Ray.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Rsensing6_khairul 1 Image Classification Image Classification uses the spectral information represented by the digital numbers in one or more spectral.
Boris Babenko Department of Computer Science and Engineering University of California, San Diego Semi-supervised and Unsupervised Feature Scaling.
Land Cover Classification Defining the pieces that make up the puzzle.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
1 Remote Sensing and Image Processing: 6 Dr. Mathias (Mat) Disney UCL Geography Office: 301, 3rd Floor, Chandler House Tel:
Summer Session 09 August Tips for the Final Exam Make sure your answers clear, without convoluted language. Read questions carefully – are you answering.
Image Classification 영상분류
Remote Sensing Supervised Image Classification. Supervised Image Classification ► An image classification procedure that requires interaction with the.
November 13, 2014Computer Vision Lecture 17: Object Recognition I 1 Today we will move on to… Object Recognition.
1 E. Fatemizadeh Statistical Pattern Recognition.
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
Map of the Great Divide Basin, Wyoming, created using a neural network and used to find likely fossil beds See:
Digital Image Processing
Change Detection The process of identifying differences in the state of an object or phenomenon by observing it at different times. Change detection applications:
Computer Graphics and Image Processing (CIS-601).
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Colour and Texture. Extract 3-D information Using Vision Extract 3-D information for performing certain tasks such as manipulation, navigation, and recognition.
Various Change Detection Analysis Techniques. Broadly Divided in Two Approaches ….. 1.Post Classification Approach. 2.Pre Classification Approach.
Unsupervised Classification
Gaussian Mixture Model classification of Multi-Color Fluorescence In Situ Hybridization (M-FISH) Images Amin Fazel 2006 Department of Computer Science.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
26. Classification Accuracy Assessment
Supervised Training and Classification
IMAGE PROCESSING RECOGNITION AND CLASSIFICATION
Unsupervised Classification in Imagine
LECTURE 10: DISCRIMINANT ANALYSIS
Map of the Great Divide Basin, Wyoming, created using a neural network and used to find likely fossil beds See:
Incorporating Ancillary Data for Classification
University College London (UCL), UK
REMOTE SENSING Multispectral Image Classification
REMOTE SENSING Multispectral Image Classification
Computer Vision Lecture 16: Texture II
Process Capability.
Unsupervised Classification
Image Information Extraction
Object Recognition Today we will move on to… April 12, 2018
EE513 Audio Signals and Systems
University College London (UCL), UK
ALI assignment – see amended instructions
LECTURE 09: DISCRIMINANT ANALYSIS
Environmental Remote Sensing GEOG 2021
Presentation transcript:

Environmental Remote Sensing Lecture 5: Image Classification " Purpose: – categorising data – data abstraction / simplification – data interpretation – mapping " for land cover mapping " use land cover class as a surrogate for other information of interest (ie assign relevant information/characteristics to a land cover class)

Basis for 'classifying' " method: pattern recognition " use any/all of the following properties in an image to differentiate between land cover classes:- – spectral – spatial – temporal – directional – [time / distance-resolved (LIDAR)]

Spatial Pattern Recognition " attempt to make use of spatial context to distinguish between different classes – e.g. measures of image texture, shape descriptors derived from image data, spatial context of 'objects ' derived from data. temporal pattern recognition " the ability to distinguish based on spectral or spatial considerations may vary over the year " use variations in image DN (or derived data) over time to distinguish between different cover types – e.g. variations in VI over agricultural crops

directional pattern recognition " surface with different structures will tend to give different trends in reflectance as a function of view and illumination angles – recent work indicates that we may be able to make use of such information to distinguish between surfaces with different structural attributes, even if they are not clearly separable on the basis of spectral properties alone. spectral pattern recognition " most typically used " relies on being able to distinguish between different land cover classes from differences in the spectral reflectance (or more typically, image DN) in a number of wavebands

Approaches to Classification " Whatever the particular feature(s) of a surface we are attempting to use to distinguish it from other surfaces, we need some form of automated classification algorithm to achieve this on a computer. – Supervised Classification – Unsupervised Classification

Supervised Classification " training stage – preprocessing of images e.g. using Principal Components analysis may be advisable sometimes to determine the intrinsic dimensionality of the data " Identify areas of cover types of interest (map, ground survey, spectral characteristics) in bands of an image – typically, a FCC image is used, with three bands which will enhance discrimination between cover types of interest – utility of FCC is dependent on display capabilities, and individual bands can also be used (or band transformations, such as NDVI which will enhance vegetation, or various principal component bands )

" areas of interest are delineated by the user " spectral information on the cover types is gathered for these areas, which can then be plotted in 'feature space' – typically, only a two-dimensional scatterplot can be displayed at one time (two bands) " the scatterplot(s) should be examined to determine the potential separability of the land cover classes

Supervised classification - classification stage " minimum distance to means – for each land cover class, calculate the mean vector in feature space (i.e. the mean value in each waveband) – segment the feature space into regions within which a pixel will lie closest to a particular class mean – we can typically define some limit beyond which a pixel is defined as unclassified " a simple and fast technique " major limitation:- – insensitive to different degrees of variance in the data (so not typically used if classes are close together or have different degrees of variance)

" parallelepiped ('box') –segment feature space by assigning boundaries around the spread of a class in feature space (taking into account variations in the variance of classes) –N-dimensional equivalent of applying (upper and lower) thresholds to a single band of data –typically use minimum/maximum of DN in a particular class to define limits, giving a rectangular 'box' for each class (in 2D) –pixels outside of these regions are unclassified (which is good or bad, depending on what you want!!) " problems if class regions overlap " problems if there is a high degree of covariance in the data between different bands (a rectangular box shape is inappropriate) – can modify algorithm by using stepped boundaries with a series of rectangles to partially overcome such problems " simple and fast technique " takes some account of variations in the variance of each class

" (Gaussian) maximum likelihood " based on the assumption that data in a class is (unimodal) Gaussian distributed (Normal) – distribution of pixels in a class can then be defined through a mean vector and covariance matrix – calculate the probability of a pixel belonging to any class using probability density functions defined from this information – we can represent this as ellipsoidal equiprobability contours – assign a pixel to the class for which it has the highest probability of belonging to – can apply a threshold where, if the probability is below this, it is assigned as unclassified " relatively computationally-intensive

Bayesian " an extension of the maximum likelihood classifier – assign two weighting factors in determining the probabilities:- " determine an a priori probability of the occurrence of a class in the scene " assign a 'cost' of misclassification for each class – combine these factors to determine the probability contours – preferable if relevant information is available

Decision tree " apply a classification in a series of steps, where the classifier has only to be able to distinguish between two or more classes at each step – can combine various types of classifiers as appropriate using such methods

application of classification stage – The selected algorithm is run on a computer accuracy assessment stage – class separation (at training stage) – useful to have some quantitative estimate of class separability in all bands – e.g. divergence " distance between class means, weighted by covariance of the classes " the larger the divergence, the larger the 'statistical distance' between classes and the higher the likelihood of correct classification

" Contingency table (confusion matrix) – classification is applied to the training set pixels – since the cover type of these pixels is 'known', we can get an indication of the degree of error (misclassification) in the algorithm used – it does not therefore provide an indication of the overall classification accuracy, rather, it tells us how well the classifier can classify the training areas – this information may be used to indicate that the user should go back and reinvestigate, retrain, or reclassify. " ideally, one should not in fact use the same pixels in the classification 'accuracy' check as in the training; an independent set of samples should be used, which can be used to give a better 'overall' accuracy estimate

Unsupervised Classification (clustering) " cluster pixels in feature space based on some measure of their proximity " interpretation of results / assigned classes –need to compare clustered classes with reference data to interpret (or, in ideal conditions, use spectral information to infer) –clusters occurring in feature space might not be those a human operator would select –can be useful, e.g. in picking up variations within what would otherwise be distinguished as a single class (e.g. stressed / unstressed fields of the same crop) –can sometimes be difficult to interpret, especially if a cluster represents data from a mixture of cover types which just happen not to be spectrally separable –the clusters can sometimes be of little intrinsic value in themselves, e.g. sunlit trees, shaded trees is perhaps not a useful discrimination if one simply wants to classify 'trees', and so clusters may have to be combined

" A large number of clustering algorithms exist " K-means – input the number of clusters desired – algorithm typically initiated with arbitrarily-located 'seeds' for cluster means – each pixel then assigned to closest cluster mean – revised mean vectors are then computed for each cluster – iterate on these two operations until some convergence criterion is met (cluster means don't move between iterations) – method is computationally-expensive, because it is iterative

Hybrid Approaches " useful when there is a large degree of variability in the DN of individual classes " use clustering concepts from unsupervised classification to derive sub- classes for individual classes, followed by standard supervised methods. " can apply e.g. K-means algorithm to (test) subareas, to derive class statistics and use the derived clusters to classify the whole scene " requirement that all classes of interest are represented in these test areas " clustering algorithms may not always determine all relevant classes in an image, and so the results may have to be combined with supervised information, e.g. linear features (roads etc.) may not be picked-up by the textural methods described above

Postclassification filtering " The result of a classification from RS data can often appear rather 'noisy' " it is sometimes desirable to aggregate local information in some way, e.g. in order to gain an overall picture of local cover type, or the most common local cover type " a variety of aggregation techniques exist, the simplest and perhaps most common of which is majority filtering –a kernel is passed over the classification result and the class which occurs most commonly in the kernel is used " Although this is the most common method, it may not always be appropriate; the particular method for spatial aggregation of categorical data of this sort depends on the particular application to which the data are to be put –e.g. successive aggregations will typically lose scattered data of a certain class, but keep tightly-clustered data