Presentation is loading. Please wait.

Presentation is loading. Please wait.

Pixel-based image classification

Similar presentations


Presentation on theme: "Pixel-based image classification"— Presentation transcript:

1 Pixel-based image classification
Lecture 8

2 What is image classification or pattern recognition
Is a process of classifying multispectral (hyperspectral) images into patterns of varying gray or assigned colors that represent either clusters of statistically different sets of multiband data, some of which can be correlated with separable classes/features/materials. This is the result of Unsupervised Classification, or numerical discriminators composed of these sets of data that have been grouped and specified by associating each with a particular class, etc. whose identity is known independently and which has representative areas (training sites) within the image where that class is located. This is the result of Supervised Classification. Spectral classes are those that are inherent in the remote sensor data and must be identified and then labeled by the analyst. Information classes are those that human beings define.

3 supervised classification
supervised classification. Identify known a priori through a combination of fieldwork, map analysis, and personal experience as training sites; the spectral characteristics of these sites are used to train the classification algorithm for eventual land-cover mapping of the remainder of the image. Every pixel both within and outside the training sites is then evaluated and assigned to the class of which it has the highest likelihood of being a member. unsupervised classification, The computer or algorithm automatically group pixels with similar spectral characteristics (means, standard deviations, covariance matrices, correlation matrices, etc.) into unique clusters according to some statistically determined criteria. The analyst then re-labels and combines the spectral clusters into information classes.

4 Hard vs. Fuzzy classification
Supervised and unsupervised classification algorithms typically use hard classification logic to produce a classification map that consists of hard, discrete categories (e.g., forest, agriculture). Conversely, it is also possible to use fuzzy set classification logic, which takes into account the heterogeneous and imprecise nature (mix pixels) of the real world. Proportion of the m classes within a pixel (e.g., 10% bare soil, 10% shrub, 80% forest). Fuzzy classification schemes are not currently standardized.

5

6 Pixel-based vs. Object-oriented classification
In the past, most digital image classification was based on processing the entire scene pixel by pixel. This is commonly referred to as per-pixel (pixel-based) classification. Object-oriented classification techniques allow the analyst to decompose the scene into many relatively homogenous image objects (referred to as patches or segments) using a multi-resolution image segmentation process. The various statistical characteristics of these homogeneous image objects in the scene are then subjected to traditional statistical or fuzzy logic classification. Object-oriented classification based on image segmentation is often used for the analysis of high-spatial-resolution imagery (e.g., 1  1 m Space Imaging IKONOS and 0.61  0.61 m Digital Globe QuickBird).

7 Knowledge-based information extraction: Artificial Intelligence
Neural network Decision tree Support vector machine (SVM)

8 Purposes of classification
Land use and land cover (LULC) Vegetation types Geologic terrains Mineral exploration Alteration mapping …….

9 Example spectral plot Two bands of data.
Each pixel marks a location in this 2d spectral space Our eye’s can split the data into clusters. Some points do not fit clusters.

10 1. Unsupervised classification
Uses statistical techniques to group n-dimensional data into their natural spectral clusters, and uses the iterative procedures label certain clusters as specific information classes K-mean and ISODATA For the first iteration arbitrary starting values (i.e., the cluster properties) have to be selected. These initial values can influence the outcome of the classification. In general, both methods assign first arbitrary initial cluster values. The second step classifies each pixel to the closest cluster. In the third step the new cluster mean vectors are calculated based on all the pixels in one cluster. The second and third steps are repeated until the "change" between the iteration is small. The "change" can be defined in several different ways, either by measuring the distances of the mean cluster vector have changed from one iteration to another or by the percentage of pixels that have changed between iterations. The ISODATA algorithm has some further refinements by splitting and merging of clusters. Clusters are merged if either the number of members (pixel) in a cluster is less than a certain threshold or if the centers of two clusters are closer than a certain threshold. Clusters are split into two different clusters if the cluster standard deviation exceeds a predefined value and the number of members (pixels) is twice the threshold for the minimum number of members.

11 Advantages Disadvantages Requires no prior knowledge of the region
Human error is minimized Unique classes are recognized as distinct units Disadvantages Classes do not necessarily match informational categories of interest Limited control of classes and identities Spectral properties of classes can change with time

12 Distance Measures are used to group or cluster brightness values together
Euclidean distance between points in space is a common way to calculate closeness

13 K-means (unsupervised)
A set number of cluster centers are positioned randomly through the spectral space. Pixels are assigned to their nearest cluster. The mean location is re-calculated for each cluster. Repeat 2 and 3 until movement of cluster centres is below threshold. Assign class types to spectral clusters.

14 Example k-means 1. First iteration. The cluster centers are set at random. Pixels will be assigned to the nearest center. 2. Second iteration. The centers move to the mean-center of all pixels in this cluster. 3. N-th iteration. The centers have stabilized.

15

16

17 Example ISODATA 1. Data is clustered but blue cluster is very stretched in band 1. 2.Cyan and green clusters only have 2 or less pixels. So they will be removed. 3. Either assign outliers to nearest cluster, or mark as unclassified.

18 ISODATA: Initial Cluster Values (properties)
number of classes maximum iterations pixel change threshold ( %) (The change threshold is used to end the iterative process when the number of pixels in each class changes by less than the threshold. The classification will end when either this threshold is met or the maximum number of iterations has been reached) initializing from statistics (Erdas) or from input (ENVI) (the initial values to put in for ENVI are minimum # pixel in class, maximum class stdv, minimum class distance, maximum # merge pairs)

19 Maximum Class Stdv (in pixel value)
Maximum Class Stdv (in pixel value). If the stdv of a class is larger than this threshold then the class is split into two classes. Minimum class distance (in pixel value) between class means. If the distance between two class means is less than the minimum value entered, then ENVI merges the classes. Optional Maximum stdev from mean (1 to 3σ) and maximum distance error (in pixel value). If any of these two setup, the some pixels might not be classified.

20 5-10 classes, 8 iterations, 5 for change threshold, (MinP 5, MaxSD 1, MinD 5, MMP 2)

21 1-5 classes, 11 iterations, 5 for change threshold, (MinP 5, MaxSD 1, MinD 5, MMP 2)

22 5 classes 10 classes

23 2. Supervised classification: training sites selection
Based on known a priori through a combination of fieldwork, map analysis, and personal experience on-screen selection of polygonal training data (ROI), and/or on-screen seeding of training data (ENVI does not have this, Erdas Imagine does). The seed program begins at a single x, y location and evaluates neighboring pixel values in all bands of interest. Using criteria specified by the analyst, the seed algorithm expands outward like an amoeba as long as it finds pixels with spectral characteristics similar to the original seed pixel. This is a very effective way of collecting homogeneous training information. From spectral library of field measurements

24 Advantages Analyst has control over the selected classes tailored to the purpose Has specific classes of known identity Does not have to match spectral categories on the final map with informational categories of interest Can detect serious errors in classification if training areas are missclassified

25 Disadvantages Analyst imposes a classification (may not be natural)
Training data are usually tied to informational categories and not spectral properties Remember diversity Training data selected may not be representative Selection of training data may be time consuming and expensive May not be able to recognize special or unique categories because they are not known or small

26 Statistic extraction of each training site
Each pixel in each training site associated with a particular class (c) is represented by a measurement vector, Xc; Average of all pixels in a training site called mean vector, Mc; a covariance matrix of Vc. where BVi,j,k is the brightness value for the i,jth pixel in band k. µck represents the mean value of all pixels obtained for class c in band k. Covckl is the covariance of class c between bands l through k.

27 Selecting ROIs Alfalfa Cotton Grass Fallow

28 Spectra of ROIs from ETM+ image

29 Spectra from library Resampled to match TM/ETM+, 6 bands

30 Supervised classification methods
Various supervised classification algorithms may be used to assign an unknown pixel to one of m possible classes. The choice of a particular classifier or decision rule depends on the nature of the input data and the desired output. Parametric classification algorithms assumes that the observed measurement vectors Xc obtained for each class in each spectral band during the training phase of the supervised classification are Gaussian; that is, they are normally distributed. Nonparametric classification algorithms make no such assumption. Several widely adopted nonparametric classification algorithms include: one-dimensional density slicing parallepiped, minimum distance, nearest-neighbor, and neural network and expert system analysis. The most widely adopted parametric classification algorithms is the: maximum likelihood. Hyperspectral classification methods Binary Encoding Spectral Angle Mapper Matched Filtering Spectral Feature Fitting Linear Spectral Unmixing

31 2.1 Parallepiped This is a widely used digital image classification decision rule based on simple Boolean “and/or” logic. If a pixel value lies above the low threshold and below the high threshold for all n bands being classified, it is assigned to that class. If the pixel value falls in multiple classes, ENVI assigns the pixel to the last class matched. Areas that do not fall within any of the parallelepipeds are designated as unclassified. In ENVI, you can use 1-3

32 This method is a computationally efficient method
This method is a computationally efficient method. But an unknown pixel might meet the criteria of more than one class and is always assigned to the first class for which it meets all criteria. The Minimum Distance to Means can assign any pixel to just one class.

33 Parallelepiped example
Training classes plotted in spectral space. In this example using 2 bands.

34 Parallelepiped example continued
Each class type defines a spectral box Note that some boxes overlap even though the classes are spatially separable. This is due to band correlation in some classes. Can be overcome by customising boxes.

35 1 means 1 stdev from mean, 2 means 2 stdev from mean, 3 means 3 stdev from mean;
Use 1, you will classify the closest pixels to the class Use 3, you will include some not so closest pixels to the class

36

37 2.2 Minimum distance The distance used in a minimum distance to means classification algorithm can take two forms: the Euclidean distance based on the Pythagorean theorem and the “round the block” distance. The Euclidean distance is more computationally intensive, but it is more frequently used All pixels are classified to the nearest class unless a standard deviation or distance threshold is specified, in which case some pixels may be unclassified if they do not meet the selected criteria. e.g. the distance of point a to class forest is

38 If either Max stdev or Max distance error is not set, all pixels will be classified.
If the Max stdev from mean is set at 2 (stdev), then the pixels with values outside the mean ± 2σ will not be classified. If the Max distance error is set at 4.2 (pixel value), then the pixels with distance larger than 4.2 will not be classified.

39 2.3 Maximum likelihood Instead based on training class multispectral distance measurements, the maximum likelihood decision rule is based on probability. The maximum likelihood procedure assumes that each training class in each band are normally distributed (Gaussian). Training data with bi- or n-modal histograms in a single band are not ideal. In such cases the individual modes probably represent unique classes that should be trained upon individually and labeled as separate training classes. the probability of a pixel belonging to each of a predefined set of m classes is calculated based on a normal probability density function, and the pixel is then assigned to the class for which the probability is the highest. probability

40 The estimated probability density function for class wi (e. g
The estimated probability density function for class wi (e.g., forest) is computed using the equation: where exp [ ] is e (the base of the natural logarithms) raised to the computed power, x is one of the brightness values on the x-axis, is the estimated mean of all the values in the forest training class, and is the estimated variance of all the measurements in this class. Therefore, we need to store only the mean and variance of each training class (e.g., forest) to compute the probability function associated with any of the individual brightness values in it.

41 For multiple bands of remote sensor data for the classes of interest, we compute an n- dimensional multivariate normal density function using: where is the determinant of the covariance matrix, is the inverse of the covariance matrix, and is the transpose of the vector The mean vectors (Mi) and covariance matrix (Vi) for each class are estimated from the training data.

42 Without Prior Probability Information:
Decide unknown measurement vector X is in class i if, and only if, pi > pj for all i and j out of 1, 2, ... m possible classes and The assign the measurement vector X of an unknown pixel to a class, the maximum likelihood decision rule computers the value pi for each class. Then it assigns the pixel to the class that has the largest value Unless you select a probability threshold (0-1), all pixels are classified. Each pixel is assigned to the class that has the highest probability

43 Probability threshold from [0, 1]
Probability threshold from [0, 1]. 0 means zero probability of similarity, 1 means 100% probability of similarity.

44 2.4 Mahalanobis Distance M-distance is similar to the Euclidian distance It is similar to the Maximum Likelihood classification but assumes all class covariances are equal and therefore is a faster method. All pixels are classified to the closest ROI class unless you specify a distance threshold, in which case some pixels may be unclassified if they do not meet the threshold (in DN number)

45 2.5 Spectral Angle Mapper

46 2.6 Spectral Feature Fitting
compare the fit of image reflectance spectra to selected reference reflectance spectra using a least-squares technique. SFF is an absorption-feature-based methodology. Both reflectance spectra should be continuum removed. A scale image is output for each reference spectrum and is a measure of absorption feature depth which is related to material abundance. The image and reference spectra are compared at each selected wavelength in a least-squares sense and the root mean square (rms) error is determined for each reference spectrum.

47 Least square tech (regression)

48 A continuum is a mathematical function used to isolate a particular absorption feature for analysis

49 Supervised classification method: Spectral Feature Fitting
Source: .gov/html/data.html


Download ppt "Pixel-based image classification"

Similar presentations


Ads by Google