Download presentation
Presentation is loading. Please wait.
1
Affordance Prediction via Learned Object Attributes Tucker Hermans James M. Rehg Aaron Bobick Computational Perception Lab School of Interactive Computing Georgia Institute of Technology
2
Motivation Determine applicable actions for an object of interest Learn this ability for previously unseen objects 1
3
Affordances Latent actions available in the environment Joint function of the agent and object Proposed by Gibson 1977 2
4
Direct Perception Affordances are directly perceived from the environment Gibson’s original model of affordance perception Direct Perception Model 3
5
Object Models Category Affordance FullCategory Affordance Chain 4 Moore, Sun, Bobick, & Rehg, IJRR 2010
6
Attribute Affordance Model Benefits of Attributes Attributes determine affordances Scale to novel object categories Give a supervisory signal not present in feature selection Attribute-Affordance Model 5
7
Attribute Affordance Model 6 Based on Lampert et. al. CVPR 09
8
Visual Features SIFT codewords extracted densely 7 … LAB color histogram Texton filter bank
9
Attributes Shape: 2D-Boxy, 3D-Boxy, cylindrical, spherical Colors: blue, red, yellow, purple, green, orange, black, white, and gray Material: cloth, ceramic, metal, paper, plastic, rubber, and wood Size: height and width (cm) Weight (kg) Total attribute feature length: 23 total elements 8
10
Attribute Classifiers Learn attribute classifiers using binary SVM and SVM regression Use multichannel χ 2 kernel 9
11
Affordance Classifiers Binary SVM with multichannel Euclidean and hamming distance kernel Train on ground truth attribute values Infer affordance using predicted attribute values 10
12
Experimental Setup 11
13
Experimental Data Six object categories: balls, books, boxes, containers, shoes, and towels 7 Affordances: rollable, pushable, gripable, liftable, traversable, caryable, dragable 375 total images 12
14
Results: Affordance Prediction Attribute-Affordance 13 Category Affordance Chain
15
Results: Affordance Prediction Category Affordance Full 14 Attribute-Affordance
16
Results: Affordance Prediction Category Affordance ChainCategory Affordance Full 15
17
Results: Affordance Prediction Attribute-AffordanceDirect Perception 16
18
Results: Affordance Prediction AttributeDPCA-FullCA-Chain Pushable74.4383.7577.5065.56 Rollable96.8797.3290.7184.14 Graspable70.0981.2573.2155.48 Liftable73.9183.9375.7167.48 Dragable72.8781.4375.0060.00 Carryable73.9183.9375.7167.48 Traversable93.3995.0090.7186.61 Total81.1285.4679.2168.57 17 Percent correctly classified
19
Results: Attribute Prediction Color PredictionMaterial Prediction 18
20
Results: Attribute Prediction Shape PredictionObject Category Prediction 19
21
Results: Novel Object Class Attribute-AffordanceDirect Perception 20 Object class “book”
22
Results: Novel Object Class Attribute-AffordanceDirect Perception 21 Object class “box”
23
Results: Novel Object Class BallsBooksBoxesContainerShoesTowels Attribute52.0339.9969.0176.2860.9753.63 DP57.9965.5867.6958.9667.8667.91 22 Percent of correctly classified affordances across all novel object categories
24
Future Work Train attribute classifiers on larger auxiliary dataset Incorporate depth sensing Combine attribute and object models Use parts as well as attributes Affordances of elements other than individual objects Attribute-Category Model 23
25
Conclusion Current dataset does not provide a diverse enough set of object classes for attributes to provide significant information transfer Attribute model restricts use of all features, unlike direct perception which has all visual features available Attribute model outperformed object models Direct perception and attribute models are comparable for small amounts of training data 24
26
Affordance Prediction via Learned Object Attributes Tucker Hermans James M. Rehg Aaron Bobick Computational Perception Lab School of Interactive Computing Georgia Institute of Technology
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.