Action-Grounded Push Affordance Bootstrapping of Unknown Objects

Slides:



Advertisements
Similar presentations
Negative Selection Algorithms at GECCO /22/2005.
Advertisements

Analysis of High-Throughput Screening Data C371 Fall 2004.
Random Forest Predrag Radenković 3237/10
電腦視覺 Computer and Robot Vision I
1 Challenge the future HON4D: Histogram of Oriented 4D Normals for Activity Recognition from Depth Sequences Omar Oreifej Zicheng Liu CVPR 2013.
Image Indexing and Retrieval using Moment Invariants Imran Ahmad School of Computer Science University of Windsor – Canada.
Uncertainty Representation. Gaussian Distribution variance Standard deviation.
The Terrapins Computer Vision Laboratory University of Maryland.
Handwritten Thai Character Recognition Using Fourier Descriptors and Robust C-Prototype Olarik Surinta Supot Nitsuwat.
B.Ramamurthy. Data Analytics (Data Science) EDA Data Intuition/ understand ing Big-data analytics StatsAlgs Discoveries / intelligence Statistical Inference.
West Virginia University
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
COMPARISON OF IMAGE ANALYSIS FOR THAI HANDWRITTEN CHARACTER RECOGNITION Olarik Surinta, chatklaw Jareanpon Department of Management Information System.
Apache Mahout. Mahout Introduction Machine Learning Clustering K-means Canopy Clustering Fuzzy K-Means Conclusion.
Some working definitions…. ‘Data Mining’ and ‘Knowledge Discovery in Databases’ (KDD) are used interchangeably Data mining = –the discovery of interesting,
Line detection Assume there is a binary image, we use F(ά,X)=0 as the parametric equation of a curve with a vector of parameters ά=[α 1, …, α m ] and X=[x.
資訊工程系智慧型系統實驗室 iLab 南台科技大學 1 A Static Hand Gesture Recognition Algorithm Using K- Mean Based Radial Basis Function Neural Network 作者 :Dipak Kumar Ghosh,
Today Ensemble Methods. Recap of the course. Classifier Fusion
Pattern Recognition April 19, 2007 Suggested Reading: Horn Chapter 14.
Handwritten Recognition with Neural Network Chatklaw Jareanpon, Olarik Surinta Mahasarakham University.
Visual Information Systems Recognition and Classification.
Robot Kinematics: Position Analysis 2.1 INTRODUCTION  Forward Kinematics: to determine where the robot ’ s hand is? (If all joint variables are known)
Sparse Bayesian Learning for Efficient Visual Tracking O. Williams, A. Blake & R. Cipolloa PAMI, Aug Presented by Yuting Qi Machine Learning Reading.
Learning Photographic Global Tonal Adjustment with a Database of Input / Output Image Pairs.
Carl Vondrick, Aditya Khosla, Tomasz Malisiewicz, Antonio Torralba Massachusetts Institute of Technology
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
Robotics Chapter 6 – Machine Vision Dr. Amit Goradia.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
Data Summit 2016 H104: Building Hadoop Applications Abhik Roy Database Technologies - Experian LinkedIn Profile:
Introduction to Transformations / Translations. By the end of this lesson, you will know… Transformations in general: A transformation is a change in.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Machine Learning Supervised Learning Classification and Regression
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Machine Learning with Spark MLlib
Unsupervised Learning of Video Representations using LSTMs
Who am I? Work in Probabilistic Machine Learning Like to teach 
A Personal Tour of Machine Learning and Its Applications
Data Mining, Neural Network and Genetic Programming
Recognizing Deformable Shapes
Recognition: Face Recognition
Basic machine learning background with Python scikit-learn
Natural Language Processing of Knee MRI Reports
Machine Learning Basics
Machine Learning Feature Creation and Selection
Machine Learning Week 1.
Fitting Curve Models to Edges
Lecture 22 Clustering (3).
Roberto Battiti, Mauro Brunato
The Open World of Micro-Videos
Ke Chen Reading: [7.3, EA], [9.1, CMB]
Neuro-Computing Lecture 4 Radial Basis Function Network
Camera Calibration Using Neural Network for Image-Based Soil Deformation Measurement Systems Zhao, Honghua Ge, Louis Civil, Architectural, and Environmental.
Matching Words with Pictures
Image Classification Painting and handwriting identification
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Introduction to Radial Basis Function Networks
Chapter 7: Transformations
Unit 1 Transformations in the Coordinate Plane
A Novel Smoke Detection Method Using Support Vector Machine
Realtime Recognition of Orchestral Instruments
Machine Learning – a Probabilistic Perspective
Unsupervised Perceptual Rewards For Imitation Learning
Unit 1 Transformations in the Coordinate Plane
Unsupervised Networks Closely related to clustering
Point Cloud Processing
Word representations David Kauchak CS158 – Fall 2016.
Image recognition.
Introduction to Artificial Intelligence Lecture 22: Computer Vision II
What is Artificial Intelligence?
Presentation transcript:

Action-Grounded Push Affordance Bootstrapping of Unknown Objects Barry Ridge & Aleš Ude Humanoid and Cognitive Robotics Lab Department of Automation, Biocybernetics & Robotics IROS Nov 4th, 2013 1

Introduction

Grounding 3-D Shape Features in Push Actions RGB image + segmented 3-D point cloud Shape primitives fitted to object parts Learn object push affordances when objects are pushed from different positions & directions. Dense 3D point clouds from modern depth sensors capture object surface shape with reasonably high accuracy. Object models can be formed dynamically by fitting simple shape primitives to object parts in such point clouds. Shape & position of object parts can be used to learn & predict affordances when objects are pushed from different positions & directions.

Point Cloud Library (PCL)

Calibrated Point Clouds & Push Trajectories

Action Coordinate Frame Object point clouds before/after interaction + hand push trajectory Point clouds + trajectory transformed to push action coordinate frame Transform object point clouds before & after interaction to push action coordinate frame. We define the push action coordinate frame w.r.t. the push contact point & direction: - Origin centred at object contact point. - Frame orientation found by fitting line to push trajectory using least squares linear regression.

Action-Grounded Shape Features Fitting Planes to Object Parts Before Interaction Fitting Planes to Object Parts After Interaction Point Clouds + Trajectory transformed to Push Action Coordinate Frame Divide object point cloud into part cells: 1 cell for overall point cloud. 2 cells by evenly splitting point cloud by its extent in X-dimension. - 2 cells by evenly splitting point cloud by its extent in Y-dimension. - 2 cells by evenly splitting point cloud by its extent in Z-dimension. Find point centroids + Fit planes in each cell. Derive 35-dimensional object shape feature vector.

Action-Grounded Shape Features

Learning Goal Cross-View Co-Occurrences Input Feature Space (Pre-Push) Output Feature Space (Post-Push)

Self-supervised Multi-view Learning Vector Quantization

Class Discovery In order to discover affordance classes in the output layer, we perform unsupervised clustering of the prototypes using the X-means algorithm to find the optimal k* clustering...

Class Prediction ...then finally, class labels may be assigned to input layer prototypes using the following labeling function: ...and input samples may then be classified by matching them to their nearest-neighbour input layer prototype.

Experiments 5 Push Types 4 Affordance Types 5 push types: push forward from centre of object, top of object, bottom of object, left of object and right of object. 4 affordance types: forward translation, forward topple, left rotation and right rotation.

Experiments 134 object push trials performed on 5 household objects in various positions and orientations on table surface. 10-Fold Incremental Cross Validation. Comparison with incremental supervised learning methods based on vector quantization, including: GLVQ (Generalized Learning Vector Quantization) GRLVQ (Generalized Relevance Learning Vector Quantization) SRNG (Supervised Relevance Neural Gas)

Results: Short-Term Training

Results: Long-Term Training

Results: Reduced Feature Set, Short-Term Training

Results: Reduced Feature Set, Long-Term Training

Conclusions Main contribution: an action grounded 3-D visual feature descriptor to be used for bootstrapping object affordances when objects are unknown. When used in combination with a self-supervised learning algorithm, this feature descriptor is effective at facilitating both affordance class discovery and prediction in an online learning setting with a number of initially unknown objects and object affordances. A feature relevance determination extension to the self-supervised algorithm was also shown to boost affordance class prediction results by emphasizing the discriminative contribution of particular features within the descriptor.