Semi-Supervised Learning with Graph Transduction Project 2 Due Nov. 7, 2012, 10pm EST Class presentation on Nov. 12, 2012.

Slides:



Advertisements
Similar presentations
. Markov Chains. 2 Dependencies along the genome In previous classes we assumed every letter in a sequence is sampled randomly from some distribution.
Advertisements

Original Figures for "Molecular Classification of Cancer: Class Discovery and Class Prediction by Gene Expression Monitoring"
MATLAB Examples. CS 1112 MATLAB Examples Find the number of positive numbers in a vector x = input( 'Enter a vector: ' ); count = 0; for ii = 1:length(x),
Fast SDP Relaxations of Graph Cut Clustering, Transduction, and Other Combinatorial Problems (JMLR 2006) Tijl De Bie and Nello Cristianini Presented by.
. Markov Chains as a Learning Tool. 2 Weather: raining today40% rain tomorrow 60% no rain tomorrow not raining today20% rain tomorrow 80% no rain tomorrow.
Shape Classification Based on Skeleton Path Similarity Xingwei Yang, Xiang Bai, Deguang Yu, and Longin Jan Latecki.
1. Markov Process 2. States 3. Transition Matrix 4. Stochastic Matrix 5. Distribution Matrix 6. Distribution Matrix for n 7. Interpretation of the Entries.
Self Taught Learning : Transfer learning from unlabeled data Presented by: Shankar B S DMML Lab Rajat Raina et al, CS, Stanford ICML 2007.
Maths for Computer Graphics
6 6.1 © 2012 Pearson Education, Inc. Orthogonality and Least Squares INNER PRODUCT, LENGTH, AND ORTHOGONALITY.
Chapter 5 Orthogonality
Mismatch string kernels for discriminative protein classification By Leslie. et.al Presented by Yan Wang.
Expectation Maximization Method Effective Image Retrieval Based on Hidden Concept Discovery in Image Database By Sanket Korgaonkar Masters Computer Science.
Cluster Analysis.  What is Cluster Analysis?  Types of Data in Cluster Analysis  A Categorization of Major Clustering Methods  Partitioning Methods.
Semi-Supervised Learning Using Randomized Mincuts Avrim Blum, John Lafferty, Raja Reddy, Mugizi Rwebangira.
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
Active Appearance Models Computer examples A. Torralba T. F. Cootes, C.J. Taylor, G. J. Edwards M. B. Stegmann.
Semi-supervised protein classification using cluster kernels Jason Weston, Christina Leslie, Eugene Ie, Dengyong Zhou, Andre Elisseeff and William Stafford.
6 6.1 © 2012 Pearson Education, Inc. Orthogonality and Least Squares INNER PRODUCT, LENGTH, AND ORTHOGONALITY.
Shape Classification Using the Inner-Distance Haibin Ling David W. Jacobs IEEE TRANSACTION ON PATTERN ANAYSIS AND MACHINE INTELLIGENCE FEBRUARY 2007.
Lesson 4 Review of Vectors and Matrices. Vectors A vector is normally expressed as or in terms of unit vectors likewise.
Arithmetic Operations on Matrices. 1. Definition of Matrix 2. Column, Row and Square Matrix 3. Addition and Subtraction of Matrices 4. Multiplying Row.
Evaluating Performance for Data Mining Techniques
Feature Matching Longin Jan Latecki. Matching example Observe that some features may be missing due to instability of feature detector, view change, or.
Random Walks and Semi-Supervised Learning Longin Jan Latecki Based on : Xiaojin Zhu. Semi-Supervised Learning with Graphs. PhD thesis. CMU-LTI ,
1 Week 3: Vectors and Matrices (Part III) READING: 2.2 – 2.4 EECS Introduction to Computing for the Physical Sciences.
Statistics and Linear Algebra (the real thing). Vector A vector is a rectangular arrangement of number in several rows and one column. A vector is denoted.
Yaomin Jin Design of Experiments Morris Method.
Vectors and Matrices In MATLAB a vector can be defined as row vector or as a column vector. A vector of length n can be visualized as matrix of size 1xn.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
A* Project Project 1 Due Sep. 26, 2012, 10pm EST Class presentation on Oct. 1, 2012.
Exploit of Online Social Networks with Community-Based Graph Semi-Supervised Learning Mingzhen Mo and Irwin King Department of Computer Science and Engineering.
ENG College of Engineering Engineering Education Innovation Center 1 Array Accessing and Strings in MATLAB Topics Covered: 1.Array addressing. 2.
ECE 576 – Power System Dynamics and Stability Prof. Tom Overbye University of Illinois at Urbana-Champaign 1 Lecture 26: Modal Analysis,
Advanced Analytics on Hadoop Spring 2014 WPI, Mohamed Eltabakh 1.
ASSESSING LEARNING ALGORITHMS Yılmaz KILIÇASLAN. Assessing the performance of the learning algorithm A learning algorithm is good if it produces hypotheses.
1 A Compact Feature Representation and Image Indexing in Content- Based Image Retrieval A presentation by Gita Das PhD Candidate 29 Nov 2005 Supervisor:
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
Meeting 18 Matrix Operations. Matrix If A is an m x n matrix - that is, a matrix with m rows and n columns – then the scalar entry in the i th row and.
1 Absolute Orientation Determination using Horn Approach.
4.1 Using Matrices Warm-up (IN) Learning Objective: to represent mathematical and real-world data in a matrix and to find sums, differences and scalar.
Support Vector Machines and Gene Function Prediction Brown et al PNAS. CS 466 Saurabh Sinha.
Applications of Supervised Learning in Bioinformatics Yen-Jen Oyang Dept. of Computer Science and Information Engineering.
Clustering Instructor: Max Welling ICS 178 Machine Learning & Data Mining.
Computational Biology Clustering Parts taken from Introduction to Data Mining by Tan, Steinbach, Kumar Lecture Slides Week 9.
Copyright © Cengage Learning. All rights reserved. 2 SYSTEMS OF LINEAR EQUATIONS AND MATRICES Warm Up Read pp up to and including the box labeled.
1 Lecture 3 Post-Graduate Students Advanced Programming (Introduction to MATLAB) Code: ENG 505 Dr. Basheer M. Nasef Computers & Systems Dept.
Stochastic Processes and Transition Probabilities D Nagesh Kumar, IISc Water Resources Planning and Management: M6L5 Stochastic Optimization.
Market analysis for the S&P500 Giulio Genovese Tuesday, December
Image Similarity Presented By: Ronak Patel Guided By: Dr. Longin Jan Latecki.
An Application of Divergence Estimation to Projection Retrieval for Semi-supervised Classification and Clustering MADALINA FITERAU ARTUR DUBRAWSKI ICML.
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
ESTIMATING WEIGHT Course: Special Topics in Remote Sensing & GIS Mirza Muhammad Waqar Contact: EXT:2257 RG712.
DATA MINING: CLUSTER ANALYSIS (3) Instructor: Dr. Chun Yu School of Statistics Jiangxi University of Finance and Economics Fall 2015.
13.4 Product of Two Matrices
Semi-Supervised Clustering
Matrices - Addition and Subtraction
Markov Chains Mixing Times Lecture 5
Multimodal Learning with Deep Boltzmann Machines
Classification with Perceptrons Reading:
CSE 4705 Artificial Intelligence
Image Retrieval Longin Jan Latecki.
CSCI N207 Data Analysis Using Spreadsheet
Section 2.4 Matrices.
Vectors and Matrices In MATLAB a vector can be defined as row vector or as a column vector. A vector of length n can be visualized as matrix of size 1xn.
Mingzhen Mo and Irwin King
Matrix Addition, C = A + B Add corresponding elements of each matrix to form elements of result matrix. Given elements of A as ai,j and elements of B as.
Classification by multivariate linear regression
Matching of Sets of Points
Presentation transcript:

Semi-Supervised Learning with Graph Transduction Project 2 Due Nov. 7, 2012, 10pm EST Class presentation on Nov. 12, 2012

Your task We are given a set of database shapes belonging to 70 shape classes. The shapes are the labeled data points, since for each shape we know its class label. We are also given a set of query shapes, and our goal is to find out for each query shape to which shape class does it belong. Hence we do not know the class labels of the query shapes, and our goal is to use Semi-Supervised Learning (SSL) with graph transduction to assign labels to them.

Test Datasest The input data is given as a 1400 x 1400 matrix A of pairwise affinities between 1400 shapes. You can download it from It represents similarities of shapes, the higher the value A(i,j) the more similar is shape i to shape j. Matrix A was computed based on the Inner Distance Shape Context (IDSC), which is the method used to compute it. The method is described in H. Ling and D. Jacobs, “Shape classification using the inner-distance,” IEEE Trans. PAMI, vol. 29(2), pp. 286–299, Some shape examples are shown on the next slide. The shapes are taken from a famous dataset called MPEG-7 Core Experiment CE-Shape-1 dataset. It contains 1400 shapes grouped into 70 classes with 20 shapes in each class. The list of classes and their shape examples can be seen on Some examples are shown on the next slide.

The first column shows the query shape. The remaining 10 columns show the most similar shapes retrieved by IDSC (odd row numbers) and by my method after graph transduction. (even row numbers)

Evaluators’ job Generate 5 sets of the test data using gentestdata.m 1.Normalize matrix A to a stochastic matrix P, i.e., sum of each row of P is one. 2.Generate matrix P LL of size l x l, where l = 70, by taking 1 random shape from each of the 70 shape classes. Hence there is u=1330 unlabeled points. 3.Divide the matrix P of size 1400 x 1400 into the following 3. Save matrices PLL, PLU, PUL, PUU, vector y= True_LL with class labels of size l x 1 corresponding to the labeled points, and vector True_UU of size u x 1 with labels of unlabeled points. Perform evaluation using testssl.m: For each of the 5 test datasets do 1. Call your function ssl.m to predict the class labels of shapes l+1 to l+u in matrix P. Input: matrices PLL, PLU, PUL, PUU and vector y with class labels. Output: vector z of predicted class labels of size u x Compare vector z to vector True_LL and compute the classification accuracy = (# of correct labels in z vector) / (length of z vector). Then average classification accuracy over the 5 tests and save it.

Your job Write a Matlab function called ssl.m: Input: matrices PLL, PLU, PUL, PUU and vector True_LL with class labels. Output: vector z of class labels of the remaining elements of size u x 1. You can assume that the matrices are obtained as describe on the previous slide.

Winning team The winning team is the one with the largest classification accuracy over the 5 runs. The run time will be used as tie-breaker.