Recognition of Fragmented Characters Using Multiple Feature-Subset Classifiers Recognition of Fragmented Characters Using Multiple Feature-Subset Classifiers.

Slides:



Advertisements
Similar presentations
Patient information extraction in digitized X-ray imagery Hsien-Huang P. Wu Department of Electrical Engineering, National Yunlin University of Science.
Advertisements

Relevant characteristics extraction from semantically unstructured data PhD title : Data mining in unstructured data Daniel I. MORARIU, MSc PhD Supervisor:
Road-Sign Detection and Recognition Based on Support Vector Machines Saturnino, Sergio et al. Yunjia Man ECG 782 Dr. Brendan.
Zhimin CaoThe Chinese University of Hong Kong Qi YinITCS, Tsinghua University Xiaoou TangShenzhen Institutes of Advanced Technology Chinese Academy of.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Face Recognition. Introduction Why we are interested in face recognition? Why we are interested in face recognition? Passport control at terminals in.
Spatial Filtering (Chapter 3)
Paper presentation for CSI5388 PENGCHENG XI Mar. 23, 2005
Watching Unlabeled Video Helps Learn New Human Actions from Very Few Labeled Snapshots Chao-Yeh Chen and Kristen Grauman University of Texas at Austin.
A Comprehensive Study on Third Order Statistical Features for Image Splicing Detection Xudong Zhao, Shilin Wang, Shenghong Li and Jianhua Li Shanghai Jiao.
São Paulo Advanced School of Computing (SP-ASC’10). São Paulo, Brazil, July 12-17, 2010 Looking at People Using Partial Least Squares William Robson Schwartz.
Computer Vision Spring ,-685 Instructor: S. Narasimhan Wean 5403 T-R 3:00pm – 4:20pm Lecture #20.
Chapter 5: Linear Discriminant Functions
0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
A Study of Approaches for Object Recognition
Implementing a reliable neuro-classifier
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
FACE RECOGNITION, EXPERIMENTS WITH RANDOM PROJECTION
Data Mining on NIJ data Sangjik Lee. Unstructured Data Mining Text Keyword Extraction Structured Data Base Data Mining Image Feature Extraction Structured.
Computer Vision I Instructor: Prof. Ko Nishino. Today How do we recognize objects in images?
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Comparing Kernel-based Learning Methods for Face Recognition Zhiguo Li
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Classification with Hyperplanes Defines a boundary between various points of data which represent examples plotted in multidimensional space according.
Identifying Computer Graphics Using HSV Model And Statistical Moments Of Characteristic Functions Xiao Cai, Yuewen Wang.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Hubert CARDOTJY- RAMELRashid-Jalal QURESHI Université François Rabelais de Tours, Laboratoire d'Informatique 64, Avenue Jean Portalis, TOURS – France.
1 Template-Based Classification Method for Chinese Character Recognition Presenter: Tienwei Tsai Department of Informaiton Management, Chihlee Institute.
1 An Efficient Classification Approach Based on Grid Code Transformation and Mask-Matching Method Presenter: Yo-Ping Huang Tatung University.
Smart RSS Aggregator A text classification problem Alban Scholer & Markus Kirsten 2005.
Presented by Tienwei Tsai July, 2005
Shawlands Academy Higher Computing Data Representation.
1 Recognition of Multi-Fonts Character in Early-Modern Printed Books Chisato Ishikawa(1), Naomi Ashida(1)*, Yurie Enomoto(1), Masami Takata(1), Tsukasa.
Local Non-Negative Matrix Factorization as a Visual Representation Tao Feng, Stan Z. Li, Heung-Yeung Shum, HongJiang Zhang 2002 IEEE Presenter : 張庭豪.
A Regression Approach to Music Emotion Recognition Yi-Hsuan Yang, Yu-Ching Lin, Ya-Fan Su, and Homer H. Chen, Fellow, IEEE IEEE TRANSACTIONS ON AUDIO,
Image Classification 영상분류
Yang, Luyu.  Postal service for sorting mails by the postal code written on the envelop  Bank system for processing checks by reading the amount of.
資訊工程系智慧型系統實驗室 iLab 南台科技大學 1 A Static Hand Gesture Recognition Algorithm Using K- Mean Based Radial Basis Function Neural Network 作者 :Dipak Kumar Ghosh,
Using Support Vector Machines to Enhance the Performance of Bayesian Face Recognition IEEE Transaction on Information Forensics and Security Zhifeng Li,
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
Ensemble with Neighbor Rules Voting Itt Romneeyangkurn, Sukree Sinthupinyo Faculty of Computer Science Thammasat University.
November 30, PATTERN RECOGNITION. November 30, TEXTURE CLASSIFICATION PROJECT Characterize each texture so as to differentiate it from one.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
SAR-ATR-MSTAR TARGET RECOGNITION FOR MULTI-ASPECT SAR IMAGES WITH FUSION STRATEGIES ASWIN KUMAR GUTTA.
COP5992 – DATA MINING TERM PROJECT RANDOM SUBSPACE METHOD + CO-TRAINING by SELIM KALAYCI.
Supervisor: Nakhmani Arie Semester: Winter 2007 Target Recognition Harmatz Isca.
1 An Efficient Classification Approach Based on Grid Code Transformation and Mask-Matching Method Presenter: Yo-Ping Huang.
GENDER AND AGE RECOGNITION FOR VIDEO ANALYTICS SOLUTION PRESENTED BY: SUBHASH REDDY JOLAPURAM.
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Speech Lab, ECE, State University of New York at Binghamton  Classification accuracies of neural network (left) and MXL (right) classifiers with various.
Hand Gesture Recognition Using Haar-Like Features and a Stochastic Context-Free Grammar IEEE 高裕凱 陳思安.
Reporter: Shau-Shiang Hung( 洪紹祥 ) Adviser:Shu-Chen Cheng( 鄭淑真 ) Date:99/06/15.
2D-LDA: A statistical linear discriminant analysis for image matrix
Machine Learning: A Brief Introduction Fu Chang Institute of Information Science Academia Sinica ext. 1819
Digital Image Processing CSC331
3D Face Recognition Using Range Images Literature Survey Joonsoo Lee 3/10/05.
Incremental Reduced Support Vector Machines Yuh-Jye Lee, Hung-Yi Lo and Su-Yun Huang National Taiwan University of Science and Technology and Institute.
1 A Statistical Matching Method in Wavelet Domain for Handwritten Character Recognition Presented by Te-Wei Chiang July, 2005.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
2. Skin - color filtering.
School of Computer Science & Engineering
Bag-of-Visual-Words Based Feature Extraction
Students: Meiling He Advisor: Prof. Brain Armstrong
Aline Martin ECE738 Project – Spring 2005
Announcements Project 2 artifacts Project 3 due Thursday night
Feature Selection Methods
Presentation transcript:

Recognition of Fragmented Characters Using Multiple Feature-Subset Classifiers Recognition of Fragmented Characters Using Multiple Feature-Subset Classifiers Institute of Information Science, Academia Sinica, Taiwan C.H. Chou, C.Y. Guo, and F. Chang Inter. Conf. Document Analysis and Recognition 2007

Introduction Recognizing fragmented characters, broken characters, in printed documents of poor printing quality. Recognizing fragmented characters, broken characters, in printed documents of poor printing quality. Complement to ordinary mending techniques. Complement to ordinary mending techniques. Using only intact characters as training samples. Using only intact characters as training samples. Multiple features apply to enhance recognition accuracy. Multiple features apply to enhance recognition accuracy. The resultant classifiers can classify both intact and fragmented characters with a high degree of accuracy. The resultant classifiers can classify both intact and fragmented characters with a high degree of accuracy.

Example from Chinese newspapers published between 1951 and (a) most severe (b) less severe (c) least severe

Feature Extraction Binary image, each pixel is represented by 1 (black) or 0 (white). Binary image, each pixel is represented by 1 (black) or 0 (white). LD (Linear Normalization + Density Feature) LD (Linear Normalization + Density Feature) Invariant to character fragmentation. Invariant to character fragmentation. LN → Reduction. LN → Reduction. Feature vector consists of 256 components, values range [0, 16]. Feature vector consists of 256 components, values range [0, 16]. ND (Nonlinear Sharp Normalization + Direction Feature) ND (Nonlinear Sharp Normalization + Direction Feature) Invariant to sharp deformation. Invariant to sharp deformation. NSN → Contour → 4 Direction map → Blurring → Reduction. NSN → Contour → 4 Direction map → Blurring → Reduction. Feature vector consists of 256 components, values range [0, 255]. Feature vector consists of 256 components, values range [0, 255].

Random Subspace Method The Random Subspace Method (RSM) consists in random selection of a certain number of subspaces from the original feature space, and train a classifier on each subspace Each set of training samples is derived from a set of feature vectors projected into a subspace. Each set of training samples is derived from a set of feature vectors projected into a subspace. Subspace Projection of ordinary feature vector to Sub- characters. Subspace Projection of ordinary feature vector to Sub- characters. Randomly select a small number of dimensions from a ordinary feature vector. Randomly select a small number of dimensions from a ordinary feature vector. The applied dimensions (w) of subspace: 32, 64, 128. The applied dimensions (w) of subspace: 32, 64, 128.

Random Subspace Method

Voting

Filter Model of Feature Selection RSM

Wrapper Model of Feature Selection

Architecture of the proposed method

An Example

Classification Methods

Experiment results

The accuracy of different classification methods Multiple classifiers outperform single classifiers. Hybrid feature always outperforms both LD and ND features. GCNNs performs higher accuracy than CARTs.

Computation time of the two classification methods.

The accuracy for three types of test documents LD outperforms ND for most severe and less severe data. LD outperforms ND for most severe and less severe data. ND is better than LD for least severe data. ND is better than LD for least severe data. Hybrid has the better accuracy than either LD or ND. Hybrid has the better accuracy than either LD or ND.

CARTs VS. GCNNs The accuracy rates of CARTs and GCNNs with incremental number of classifiers and different w of subspace The accuracy rates of CARTs and GCNNs with incremental number of classifiers and different w of subspace The more classifiers get the better accuracy. The more classifiers get the better accuracy. GCNNs require fewer classifiers to archive saturation accuracy than CARTs. GCNNs require fewer classifiers to archive saturation accuracy than CARTs.

CARTs VS. GCNNs

Conclusion Proposing a learning approach to deal with both intact and fragmented characters in archived newspapers The multiple predictors achieve much higher accuracy rates than single classifiers. The hybrid predictors, which use both types of feature, perform better than those using only a single feature. GCNN rule achieve higher accuracy, and require fewer classifiers, than those generated by the CART algorithm.