Presentation is loading. Please wait.

Presentation is loading. Please wait.

Group Norm for Learning Latent Structural SVMs

Similar presentations


Presentation on theme: "Group Norm for Learning Latent Structural SVMs"— Presentation transcript:

1 Group Norm for Learning Latent Structural SVMs
Daozheng Chen (UMD, College Park), Dhruv Batra (TTI-Chicago), Bill Freeman (MIT), Micah K. Johnson (GelSight, Inc.) Overview Induce Group Norm Data with complete annotation is rarely ever available. Latent variable models capture interaction between observed data (e.g. gradient histogram image features) latent or hidden variables not observed in the training data (e.g. location of object parts). Parameter estimation involve a difficult non-convex optimization problem (EM, CCCP, self-paced learning) Our goal Estimate model parameters Learn the complexity of latent variable space. Our approach norm for regularization to estimate the parameters of a latent-variable model. Key Contribution Inducing Group Norm w partitioned into P groups; each group corresponds to the parameters of a latent variable state Felzenszwalb et al. car model on the PASCAL VOC 2007 data. Each row is a component of the model. Root filters Part filters Part displacement Component #1 Component #2 norm for regularization Digit Recognition Rotation (Latent Var.) Feature Vector Images Latent Structural SVM Label space Latent Space Joint feature vector Prediction Rule: Learning objective: At group level, the norm behave like norm and induces group sparsity. Within each group, the norm behave like norm and does not promote sparsity. Alternating Coordinate and Subgradient Descent Experiment Digit recognition experiment (following the setup of Kumar et al. NIPS ‘10) MNIST data: binary classification on four difficult digit pairs (1,7), (2,7), (3,8), (8,9) Training data 5, ,742, and testing data ,135 Rotate digit images with angles from -60o to 60o PCA to form 10 dimensional feature vector Rewrite Learning Objective nonconvex convex -60o -48o -36o -24o -12o 0o 12o 24o 36o 48o 60o convex Minimize Upper bound of convex if {hi} is fixed l2 norm of the parameter vectors for different angles over the 4 digit pairs. Select only a few angles, much fewer than 22 angles Angles Not Selected -60o -48o -12o 0o -36o Subgradient Significantly higher accuracy than random sampling. 66% faster than full model with no loss in accuracy!


Download ppt "Group Norm for Learning Latent Structural SVMs"

Similar presentations


Ads by Google