Level-0 FAUST for Satlog(landsat) is from a small section (82 rows, 100 cols) of a Landsat image: 6435 rows, 2000 are Tst, 4435 are Trn. Each row is center.

Slides:



Advertisements
Similar presentations
CS 450: COMPUTER GRAPHICS LINEAR ALGEBRA REVIEW SPRING 2015 DR. MICHAEL J. REALE.
Advertisements

Information Extraction Lecture 7 – Linear Models (Basic Machine Learning) CIS, LMU München Winter Semester Dr. Alexander Fraser, CIS.
Microsoft Excel Computers Week 4.
Support Vector Machines Instructor Max Welling ICS273A UCIrvine.
SVM—Support Vector Machines
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Image Indexing and Retrieval using Moment Invariants Imran Ahmad School of Computer Science University of Windsor – Canada.
Reliability of Disk Systems. Reliability So far, we looked at ways to improve the performance of disk systems. Next, we will look at ways to improve the.
TEKS 8.6 (A,B) & 8.7 (A,D) This slide is meant to be a title page for the whole presentation and not an actual slide. 8.6 (A) Generate similar shapes using.
Support Vector Machines and Kernel Methods
Image Analysis Preprocessing Image Quantization Binary Image Analysis
Margins, support vectors, and linear programming Thanks to Terran Lane and S. Dreiseitl.
CS223B Assignment 1 Recap. Lots of Solutions! 37 Groups Many different approaches Let’s take a peek at all 37 results on one image from the test set.
Preprocessing ROI Image Geometry
CS 376 Introduction to Computer Graphics 02 / 26 / 2007 Instructor: Michael Eckmann.
Now Playing: Gong Sigur Rós From Takk... Released September 13, 2005.
The Normal Distribution
Linear Methods, cont’d; SVMs intro. Straw poll Which would you rather do first? Unsupervised learning Clustering Structure of data Scientific discovery.
SVMs Reprised Reading: Bishop, Sec 4.1.1, 6.0, 6.1, 7.0, 7.1.
Support Vector Machines
A Novel 2D To 3D Image Technique Based On Object- Oriented Conversion.
Support Vector Machines a.k.a, Whirlwind o’ Vector Algebra Sec. 6.3 SVM Tutorial by C. Burges (on class “resources” page)
By: Suhas Navada and Antony Jacob
HAWKES LEARNING SYSTEMS math courseware specialists Copyright © 2011 Hawkes Learning Systems. All rights reserved. Hawkes Learning Systems College Algebra.
Image processing Lecture 4.
Pairwise Alignment, Part I Constructing the Values and Directions Tables from 2 related DNA (or Protein) Sequences.
1. Introduction Motion Segmentation The Affine Motion Model Contour Extraction & Shape Estimation Recursive Shape Estimation & Motion Estimation Occlusion.
Translations Translations and Getting Ready for Reflections by Graphing Horizontal and Vertical Lines.
Support Vector Machines Reading: Ben-Hur and Weston, “A User’s Guide to Support Vector Machines” (linked from class web page)
Faust is fast… takes about 15 seconds on the same dataset that takes over 9 hours with knn, 40 minutes with pTree knn. I’m ready to take on oblique, need.
2009/9 1 Matrices(§3.8)  A matrix is a rectangular array of objects (usually numbers).  An m  n (“m by n”) matrix has exactly m horizontal rows, and.
Change Detection The process of identifying differences in the state of an object or phenomenon by observing it at different times. Change detection applications:
TEMPLATE DESIGN © Predicate-Tree based Pretty Good Protection of Data William Perrizo, Arjun G. Roy Department of Computer.
Support Vector Machines and Kernel Methods Machine Learning March 25, 2010.
CSC508 Convolution Operators. CSC508 Convolution Arguably the most fundamental operation of computer vision It’s a neighborhood operator –Similar to the.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 5 Describing Distributions Numerically.
1 p1 p2 p7 2 p3 p5 p8 3 p4 p6 p9 4 pa pf 9 pb a pc b pd pe c d e f a b c d e f X x1 x2 p1 1 1 p2 3 1 p3 2 2 p4 3 3 p5 6 2 p6.
Given k, k-means clustering is implemented in 4 steps, assumes the clustering criteria is to maximize intra- cluster similarity and minimize inter-cluster.
Class 26: Question 1 1.An orthogonal basis for A 2.An orthogonal basis for the column space of A 3.An orthogonal basis for the row space of A 4.An orthogonal.
CS 376b Introduction to Computer Vision 02 / 11 / 2008 Instructor: Michael Eckmann.
Graphing a Linear Inequality
6-hop myrrh example (from Damian). Market agency targeting advertising to friends of customers: Entities: 1. advertisements 2. markets 3. merchants 4.
1 Arrays of Arrays An array can represent a collection of any type of object - including other arrays! The world is filled with examples Monthly magazine:
Chapter 2: Frequency Distributions. Frequency Distributions After collecting data, the first task for a researcher is to organize and simplify the data.
Overview Data Mining - classification and clustering
Basic Communication Operations Carl Tropper Department of Computer Science.
24 October 2002Data Mining & Visualization1 Data Mining and Visualization Jeremy Walton NAG Ltd, Oxford.
5(I,C) (I,C) (I,C) (I,C)
Reliability of Disk Systems. Reliability So far, we looked at ways to improve the performance of disk systems. Next, we will look at ways to improve the.
Near Neighbor Classifiers and FAUST Faust is really a Near Neighbor Classifier (NNC) in which, for each class, we construct a big box neighborhood (bbn)
Linear Algebra Review.
Matrices Rules & Operations.
But it's pure0 so this branch ends
pTrees predicate Tree technologies
Mean Shift Segmentation
CS 4/527: Artificial Intelligence
PTrees (predicate Trees) fast, accurate , DM-ready horizontal processing of compressed, vertical data structures Project onto each attribute (4 files)
Fitting Curve Models to Edges
ID1050– Quantitative & Qualitative Reasoning
From: Perrizo, William Sent: Thursday, February 02, :45 AM To: 'Mark Silverman' The Satlog (Landsat Satellite) data set from UCI Machine Learning.
CS Fall 2016 (Shavlik©), Lecture 2
FAUST Using std or rankK instead of just gap size to determine the best gap and/or using multiple attribute cutting can improve accuracy. We need a pTree.
Functional Analytic Unsupervised and Supervised data mining Technology
Transformation Geometry
pTree-k-means-classification-sequential (pkmc-s)
FAUST{pms,std} (FAUST{pms} using # gap std
PAj>c=Pj,m om...ok+1Pj,k oi is AND iff bi=1, k is rightmost bit position with bit-value "0", ops are right binding. c = bm ...
SVMs for Document Ranking
FAUST{pms,std} (FAUST{pms} using # gap std
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Presentation transcript:

Level-0 FAUST for Satlog(landsat) is from a small section (82 rows, 100 cols) of a Landsat image: 6435 rows, 2000 are Tst, 4435 are Trn. Each row is center pixel of 3x3 cell. There are 36 8-bit feature cols - R,G,IR1,IR2 for the 9 pixels. "Class label" is the class of the central pixel. The classes are: 1=red soil 2=cotton crop 3=grey soil4=damp grey soil 5=soil w veg stubble6=mixture7=very damp grey soil R G IR1 IR2 means R G IR1 IR2 stds R G IR1 IR2 mn vals R G IR1 IR2 std vals R cls gap- std cutpt G cls gap- std cutpt IR1 cls gap- std cutpt IR2 cls gap- std cutpt % of std sum < gap? % of std sum < gap? gap- std cutpt gap- std cutpt gap- std cutpt gap- std cutpt Class differentiated-> R G ir ir % of std sum < gap? % of std sum < gap? So the criteria is: if ( 98.15< ir2 ) then class=2 else if ( 82.90< R ) then class=3 else if ( 72.83< R < 82.90) then class=4 else if ( 51.54< G < 71.36) then class=5 else if ( 71.36< G < 84.18) then class=7 else if ( 61 < R < 66.6 ) then class=1 else no class (class=0). These cut points are: mean 1 +gap*( std 1 / (std 1 +std 2 ) )

Using level-1 50% Total 1's 2's 3's 4's 5's 7's True Positives: Class Totals-> %TPs: 63.83% 43.82% 81.7% 79.1% 48.8% 66.24% 70.21% False Positives: %FPs: 19.3% 3% 0.5% 10.6% 48.8% 15.2% 40.2% Level-1 gt50% R cls g- std %stds_in_gp cutpt % % % % % G cls g- std % % % % % IR1 cls g- std % % % % % IR2 cls g- std % % % % % Using level-0 Total 1's 2's 3's 4's 5's 7's True Positives: Class Totals-> %TPs: 57.75% 21.48% 86.16% 81.86% 61.61% 63.71% 54.68% The level-0 FAUST criteria: if ( 98.15< ir2 ) then class=2 else if ( 82.90< R ) then class=3 else if ( 72.83< R < 82.90) then class=4 else if ( 51.54< G < 71.36) then class=5 else if ( 71.36< G < 84.18) then class=7 else if ( 61 < R < 66.6 ) then class=1 else no class (class=0). CLASSES: 1=red soil 2=cotton crop 3=grey soil 4=damp grey soil 5=soil w veg stubble 7=very damp grey soil The level-1 gt50% FAUST Criteria: For 3: R(83.4,inf) For 2: G(0,44.5) For 5: G(44.5,65.5) For 7: G(65.5,82.1) For 4: G(82.1,inf) ir1(88.8,100.7) For 1: ir1(112.5,inf)

ANDing the two pTrees masks the region (which is r) From notes: A Multi-attribute EIN Oblique (EINO) based heuristic: Instead of finding the best D, take the vector connecting a class mean to another class means as D To separate r from v: D=(m r  m v ) and a=|m r  m v |/2 r r r v v r m r r v v v r r v m v v r b v v r b b v b m b b b b b b b P X o (m r  m b )>|m r  m b |/2 P X o (m r  m v )>|m r  m v |/2 masks vectors that makes a shadow on m r side of the midpt r r r v v r m r r v v v r r v m v v r b v v r b b v b m b b b b b b b For classes r and b By "outermost, I mean the "furthest points away from the means in each class (in terms of their projections of the D-line); By "outermost non-outlie" I mean the furthest non-outlier points; Other possibilities: the best rankK points, the best std points, etc. Comments on where to go from here (assuming we can do the above): I think the "medoid-to-mediod" method on this page is close to optimal provided the classes are convex. If they are not convex, then some sort of Support Vector Machines, SVMs, would be the next step. In SVMs the space is translated to higher dimensions in such a way that the classes ARE convex. The inner product in that space is equivalent to a kernel function in the original space so that one need not even do the translation to get inner product based results (the genius of the method). Final note: I should say "linearly separable instead of convex (slightly weaker condition). To separate r from b: D=(m r  m b ) and a=|m r  m b |/2 Question: What's the best as cutpt? mean, vector_of_medians, outermost, outermost_non-outlier? Mistake! d=D/|D|, a=(m r +m v )/2 o d Devastating to accuracy!

E.g.,? Let D=vector connecting class means and d= D/|D| P X dot d>a = P  d i X i >a FAUST-Oblique: Create tbl, TBL(class i, class j, medoid_vector i, medoid_vector j ). Notes: If we just pick the one class which when paired with r, gives max gap, then we can use max gap or max_std_Int_pt instead of max_gap_midpt. Then need std j (or variance j ) in TBL. 4. FAUST Oblique: length, std, rkK for selecting best gap and multiple attrs. formula: P (X dot D)>a X any set of vectors. D=oblique vector (Note: if D=e i, P X i > a ). AND 2 pTrees masks To separate r from v: D = (m v  m r ), a = (m v +m r )/2 o d NOTE:!!! The picture on this page could be misleading. See next slide for a clearer picture P (m b  m r )oX>(m r +m ) |/2 o d P (m v  m r )oX>(m r +m v )/2 o d masks vectors that makes a shadow on m r side of the midpt r r r v v r m r r v v v r r v m v v r b v v r b b v b m b b b b b b b For classes r and b "outermost = "furthest from means (their projs of D-line); best rankK points, best std points, etc. "medoid-to-mediod" close to optimal provided classes are convex. Best cutpoint? mean, vector_of_medians, outmost, outmost_non-outlier? D r g b grb grb grb grb grb grb grb grb grb In higher dims same (If "convex" clustered classes, FAUST{div,oblique_gap} finds them. bgr bgr r r r v v r m r r v v v r r v m v v r v v r v P (m r  m v )/|m r  m v |oX<a For classes r and v D = m r  m v a

r r r v v r m r r v v v r r v m v v r v v r v P X dot d>a = P  d i X i >a 4. FAUST Oblique: midpt, std, rkK for selecting best gap and multiple attrs. formula: P (X dot D)>a X any set of vectors. D≡ m r  m v is the oblique vector (Note: if D=e i, P X i >a ) and let d=D/|D| To separate r from v: Using means_midpoint, calculate a as follows: a Viewing m r and m v as vectors ( e.g., m r ≡origin  point_m r ), a = ( m r + (m v -m r )/2 ) o d = (m r +m v )/2 o d d

r r r v v r m r r v v v r r v m v v r v v r v P v o d>a = P  d i X i >a 4. FAUST Oblique: X any set of vectors. d=(m v -m r )/|m v -m r | To separate r from v using midpts: a What happens when we use the previous (mistaken) a = |m v -m r |/2 ? d a Cut line m v -m r all r o d are > a so all r s are classified incorrectly as v s P v o d>a

Oblique FAUST (level-0 case): R G ir1 ir2 means R G irR1 ir2 stds d a Using level-0 (Oblique without eliminating classes as they are predicted) Total 1's 2's 3's 4's 5's 7's True Positives: 204 False Positives: 64 all 5's

APPENDIX: impure pTrees (i.e., w predicate, 50%ones). The training set was ordered by class (all setosa's came first, then all versicolor then all virginica) so that level_1 pTrees could be chosen not to span classes much. Take an images as another example. If the classes are RedCars, GreenCars, BlueCars, ParkingLot, Grass, Trees, etc., and if Peano ordering is used, what if a class spans Peano squares completely? We now create pTrees from many different predicates. Should we created pTreeSets for many different orderings as well? This would be a one time expense. It would consume much more space, but space is not an issue. With more pTrees, our PGP-D protection scheme would automatically be more secure. So move the first column values to the far right for the 1 st additional Peano pTreeSet: Move the 1 st 2 columns to the right for 2 nd Peano pTreeSet, 1st 3 for 3 rd Peano pTreeSet.. For each of the added pTreeSets, same move vertically, e.g., the 25 th would be (starting with the 4 th horizontal, directly above). Move the last column to the left for the 4 th, the last 2 left for the 5 th, the last 3 left for the 6 th additional Peano pTreeSet. For each of these 6

Vertical expansions of 2 nd added pTreeSet (13 th, 14 th added pTreeSets, resp.?) GreenCar is in level_2 pix if level_2 stride=16 (level_1 stride= 4). How are training classes in Aurora? Given set of pixs for GreenCars? Or anything shape related to identify? If only pix vals for GreenCar, have to rely on indiv pix reflectances? Analyze each pixel for GreenCar. Then wouldn't benefit except might data mine GreenCars w level_2 only? Left move 3 same as right move 1 (left 2 same as rt 2...) Thus, 4 2 =16 orderings (not 64) at level-2; 4 1 =4 level-1; 4 n at level-n. Upper right corner can be in any cell in a level-n pixel and there are 4 n such cells. Always create pure1, pure0. GTE50%, = 3*4 n separate PTreeSets. Then the question is how to order pixels in a left (or up) shift? We could actually shift and then use the usual Peano? Or could keep each cell ordering the same (see below). Do shifting at level-0. Percolate it upward. Wouldn't store shifted level-0 PTreeSets since same pixelization. Construct shifted level-n pixelizations (n>0) concurrent by, one at a time, all level-0 pixel shifts (creating an additional PTreeSet only when it is a new pixelization (e.g., only first level-0 pixel shift produces new at level-1; only 1st 3 at level-2, etc.