Classification of Clothing using Interactive Perception BRYAN WILLIMON, STAN BIRCHFIELD AND IAN WALKER CLEMSON UNIVERSITY CLEMSON, SC USA ABSTRACT ISOLATION.

Slides:



Advertisements
Similar presentations
FEATURE PERFORMANCE COMPARISON FEATURE PERFORMANCE COMPARISON y SC is a training set of k-dimensional observations with labels S and C b C is a parameter.
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Object Recognition with Features Inspired by Visual Cortex T. Serre, L. Wolf, T. Poggio Presented by Andrew C. Gallagher Jan. 25, 2007.
Foreground Focus: Finding Meaningful Features in Unlabeled Images Yong Jae Lee and Kristen Grauman University of Texas at Austin.
Robust statistical method for background extraction in image segmentation Doug Keen March 29, 2001.
Bryan Willimon, Steven Hickson, Ian Walker, and Stan Birchfield IROS 2012 Vila Moura, Algarve An Energy Minimization Approach to 3D Non- Rigid Deformable.
Extracting Minimalistic Corridor Geometry from Low-Resolution Images Yinxiao Li, Vidya, N. Murali, and Stanley T. Birchfield Department of Electrical and.
INTRODUCTION Heesoo Myeong, Ju Yong Chang, and Kyoung Mu Lee Department of EECS, ASRI, Seoul National University, Seoul, Korea Learning.
Database-Based Hand Pose Estimation CSE 6367 – Computer Vision Vassilis Athitsos University of Texas at Arlington.
Contactless and Pose Invariant Biometric Identification Using Hand Surface Vivek Kanhangad, Ajay Kumar, Senior Member, IEEE, and David Zhang, Fellow, IEEE.
Texture Segmentation Based on Voting of Blocks, Bayesian Flooding and Region Merging C. Panagiotakis (1), I. Grinias (2) and G. Tziritas (3)
Bryan Willimon Master’s Thesis Defense Interactive Perception for Cluttered Environments.
Computer Vision Spring ,-685 Instructor: S. Narasimhan Wean 5403 T-R 3:00pm – 4:20pm Lecture #20.
A Study of Approaches for Object Recognition
Contents Description of the big picture Theoretical background on this work The Algorithm Examples.
Region-Level Motion- Based Background Modeling and Subtraction Using MRFs Shih-Shinh Huang Li-Chen Fu Pei-Yung Hsiao 2007 IEEE.
Human Posture Recognition with Convex Programming Hao Jiang, Ze-Nian Li and Mark S. Drew School of Computing Science Simon Fraser University Burnaby, BC,
3D Hand Pose Estimation by Finding Appearance-Based Matches in a Large Database of Training Views
Chamfer Matching & Hausdorff Distance Presented by Ankur Datta Slides Courtesy Mark Bouts Arasanathan Thayananthan.
A Novel 2D To 3D Image Technique Based On Object- Oriented Conversion.
December 2, 2014Computer Vision Lecture 21: Image Understanding 1 Today’s topic is.. Image Understanding.
Multiple Object Class Detection with a Generative Model K. Mikolajczyk, B. Leibe and B. Schiele Carolina Galleguillos.
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
Speaker: Chi-Yu Hsu Advisor: Prof. Jian-Jung Ding Leveraging Stereopsis for Saliency Analysis, CVPR 2012.
Introduction --Classification Shape ContourRegion Structural Syntactic Graph Tree Model-driven Data-driven Perimeter Compactness Eccentricity.
Distinctive Image Features from Scale-Invariant Keypoints By David G. Lowe, University of British Columbia Presented by: Tim Havinga, Joël van Neerbos.
GM-Carnegie Mellon Autonomous Driving CRL TitleAutomated Image Analysis for Robust Detection of Curbs Thrust AreaPerception Project LeadDavid Wettergreen,
Efficient Algorithms for Robust Feature Matching Mount, Netanyahu and Le Moigne November 7, 2000 Presented by Doe-Wan Kim.
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
A Method for Hand Gesture Recognition Jaya Shukla Department of Computer Science Shiv Nadar University Gautam Budh Nagar, India Ashutosh Dwivedi.
Digital Image Processing & Analysis Spring Definitions Image Processing Image Analysis (Image Understanding) Computer Vision Low Level Processes:
MESA LAB Two papers in icfda14 Guimei Zhang MESA LAB MESA (Mechatronics, Embedded Systems and Automation) LAB School of Engineering, University of California,
Digital Image Processing CCS331 Relationships of Pixel 1.
Automatic Minirhizotron Root Image Analysis Using Two-Dimensional Matched Filtering and Local Entropy Thresholding Presented by Guang Zeng.
Automatic Image Annotation by Using Concept-Sensitive Salient Objects for Image Content Representation Jianping Fan, Yuli Gao, Hangzai Luo, Guangyou Xu.
Bryan Willimon, Steven Hickson, Ian Walker, and Stan Birchfield Clemson University IROS Vilamoura, Portugal An Energy Minimization Approach to 3D.
Bryan Willimon, Stan Birchfield, Ian Walker Department of Electrical and Computer Engineering Clemson University IROS 2010 Rigid and Non-Rigid Classification.
PSEUDO-RELEVANCE FEEDBACK FOR MULTIMEDIA RETRIEVAL Seo Seok Jun.
Content-Based Image Retrieval Using Fuzzy Cognition Concepts Presented by Tienwei Tsai Department of Computer Science and Engineering Tatung University.
Vehicle Segmentation and Tracking From a Low-Angle Off-Axis Camera Neeraj K. Kanhere Committee members Dr. Stanley Birchfield Dr. Robert Schalkoff Dr.
Dengsheng Zhang and Melissa Chen Yi Lim
The Implementation of Markerless Image-based 3D Features Tracking System Lu Zhang Feb. 15, 2005.
2005/12/021 Fast Image Retrieval Using Low Frequency DCT Coefficients Dept. of Computer Engineering Tatung University Presenter: Yo-Ping Huang ( 黃有評 )
Image-Based Segmentation of Indoor Corridor Floors for a Mobile Robot
Image-Based Segmentation of Indoor Corridor Floors for a Mobile Robot Yinxiao Li and Stanley T. Birchfield The Holcombe Department of Electrical and Computer.
Design of PCA and SVM based face recognition system for intelligent robots Department of Electrical Engineering, Southern Taiwan University, Tainan County,
Bryan Willimon IROS 2011 San Francisco, California Model for Unfolding Laundry using Interactive Perception.
Vision and Obstacle Avoidance In Cartesian Space.
Non-Ideal Iris Segmentation Using Graph Cuts
By Pushpita Biswas Under the guidance of Prof. S.Mukhopadhyay and Prof. P.K.Biswas.
Journal of Visual Communication and Image Representation
An Image Retrieval Approach Based on Dominant Wavelet Features Presented by Te-Wei Chiang 2006/4/1.
Person Following with a Mobile Robot Using Binocular Feature-Based Tracking Zhichao Chen and Stanley T. Birchfield Dept. of Electrical and Computer Engineering.
Content-Based Image Retrieval Using Color Space Transformation and Wavelet Transform Presented by Tienwei Tsai Department of Information Management Chihlee.
Cell Segmentation in Microscopy Imagery Using a Bag of Local Bayesian Classifiers Zhaozheng Yin RI/CMU, Fall 2009.
Image segmentation.
A Plane-Based Approach to Mondrian Stereo Matching
DIGITAL SIGNAL PROCESSING
+ SLAM with SIFT Se, Lowe, and Little Presented by Matt Loper
SoC and FPGA Oriented High-quality Stereo Vision System
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
ISOMAP TRACKING WITH PARTICLE FILTERING
Vehicle Segmentation and Tracking in the Presence of Occlusions
Vehicle Segmentation and Tracking from a Low-Angle Off-Axis Camera
Spatiograms Versus Histograms for Region-Based Tracking
Radio Propagation Simulation Based on Automatic 3D Environment Reconstruction D. He A novel method to simulate radio propagation is presented. The method.
Binocular Stereo Vision
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
CS4670: Intro to Computer Vision
“Traditional” image segmentation
Presentation transcript:

Classification of Clothing using Interactive Perception BRYAN WILLIMON, STAN BIRCHFIELD AND IAN WALKER CLEMSON UNIVERSITY CLEMSON, SC USA ABSTRACT ISOLATION AND CLASSIFICATION FRAMEWORK ISOLATION CONCLUSION  We have proposed an approach to interactive perception in which a pile of laundry is sifted by an autonomous robot system in order to classify and label each item.  For Combo A, the average classification rate using a single image is 62.83%, while the average classification rate using all 20 images is 100%.  These results show that, on average, classification rates using robot interaction are 59% higher than those that do not use interaction. We present a system for automatically extracting and classifying items in a pile of laundry. Using only visual sensors, the robot identifies and extracts items sequentially from the pile. When an item has been removed and isolated, a model is captured of the shape and appearance of the object, which is then compared against a database of known items. The classification procedure relies upon silhouettes, edges, and other low-level image measurements of the articles of clothing. The contributions of this paper are a novel method for extracting articles of clothing from a pile of laundry and a novel method of classifying clothing using interactive perception. Experiments demonstrate the ability of the system to efficiently classify and label into one of six categories (pants, shorts, short-sleeve shirt, long-sleeve shirt, socks, or underwear). These results show that, on average, classification rates using robot interaction are 59% higher than those that do not use interaction. EXPERIMENTAL RESULTS CLASSIFICATION Isolating an individual article of clothing involves identifying and extracting an item from the pile, one at a time, without disturbing the rest of the clothes in the pile. Classifying an item requires using visual-based shape and appearance information to classify the item into one of several prespecified categories (pants, shorts, short-sleeve shirt, long-sleeve shirt, socks, or underwear). To isolate an item from the pile, an overhead image is first segmented, and the closest foreground segment (measured using stereo disparity) is selected. Chamfering is used to determine the grasp point which is then used to extract the item from the pile in an automatic way using interactive perception. Let be a query images (either frontal or side), and let be an image in the database. These images are compared using four different features to yield a match score: where N = 4 is the number of features,, and the features are given by, the absolute difference in area between the two silhouettes, where is the number of pixels in the query binary silhouette, and similarly for ;, the absolute difference in eccentricity between the two silhouettes, where is the eccentricity of the query binary silhouette, and similarly for ;, the Hausdorff distance between the edges of the two binary silhouettes;, the Hausdorff distance between the Canny edges of the original grayscale images. To ensure proper weighting, each value is the 95 th percentile of among all of the images in the database (robust maximum). Results of using 1 image for category classification (left image). results of using 20 images for category classification (right image). The extraction and isolation process: the image taken by one of the downward-facing stereo cameras, the result of graph-based segmentation, the object found along with its grasp point (red dot), the image taken by the side-facing camera, and the binary silhouettes of the front and side views of the isolated object.