LYU0203 Smart Traveller with Visual Translator for OCR and Face Recognition Supervised by Prof. LYU, Rung Tsong Michael Prepared by: Wong Chi Hang Tsang.

Slides:



Advertisements
Similar presentations
Patient information extraction in digitized X-ray imagery Hsien-Huang P. Wu Department of Electrical Engineering, National Yunlin University of Science.
Advertisements

QR Code Recognition Based On Image Processing
Road-Sign Detection and Recognition Based on Support Vector Machines Saturnino, Sergio et al. Yunjia Man ECG 782 Dr. Brendan.
Zhimin CaoThe Chinese University of Hong Kong Qi YinITCS, Tsinghua University Xiaoou TangShenzhen Institutes of Advanced Technology Chinese Academy of.
Wen-Hung Liao Department of Computer Science National Chengchi University November 27, 2008 Estimation of Skin Color Range Using Achromatic Features.
Prénom Nom Document Analysis: Document Image Processing Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
IntroductionIntroduction AbstractAbstract AUTOMATIC LICENSE PLATE LOCATION AND RECOGNITION ALGORITHM FOR COLOR IMAGES Kerem Ozkan, Mustafa C. Demir, Buket.
Simple Face Detection system Ali Arab Sharif university of tech. Fall 2012.
Image Processing IB Paper 8 – Part A Ognjen Arandjelović Ognjen Arandjelović
Image Segmentation Image segmentation (segmentace obrazu) –division or separation of the image into segments (connected regions) of similar properties.
COLORCOLOR A SET OF CODES GENERATED BY THE BRAİN How do you quantify? How do you use?
Facial feature localization Presented by: Harvest Jang Spring 2002.
Color Image Processing
Localization of Piled Boxes by Means of the Hough Transform Dimitrios Katsoulas Institute for Pattern Recognition and Image Processing University of Freiburg.
Virtual Dart: An Augmented Reality Game on Mobile Device Supervisor: Professor Michael R. Lyu Prepared by: Lai Chung Sum Siu Ho Tung.
ICIP 2000, Vancouver, Canada IVML, ECE, NTUA Face Detection: Is it only for Face Recognition?  A few years earlier  Face Detection Face Recognition 
EE 7730 Image Segmentation.
Real-time Embedded Face Recognition for Smart Home Fei Zuo, Student Member, IEEE, Peter H. N. de With, Senior Member, IEEE.
Supervised by Prof. LYU, Rung Tsong Michael Department of Computer Science & Engineering The Chinese University of Hong Kong Prepared by: Chan Pik Wah,
Processing Digital Images. Filtering Analysis –Recognition Transmission.
Segmentation Divide the image into segments. Each segment:
Face Detection: a Survey Speaker: Mine-Quan Jing National Chiao Tung University.
LYU0203 Smart Traveller with Visual Translator for OCR and Face Recognition Supervised by Prof. LYU, Rung Tsong Michael Prepared by: Wong Chi Hang Tsang.
An Approach to Korean License Plate Recognition Based on Vertical Edge Matching Mei Yu and Yong Deak Kim Ajou University Suwon, , Korea 指導教授 張元翔.
LYU 0102 : XML for Interoperable Digital Video Library Recent years, rapid increase in the usage of multimedia information, Recent years, rapid increase.
Iris localization algorithm based on geometrical features of cow eyes Menglu Zhang Institute of Systems Engineering
Smart Traveller with Visual Translator. What is Smart Traveller? Mobile Device which is convenience for a traveller to carry Mobile Device which is convenience.
A Novel 2D To 3D Image Technique Based On Object- Oriented Conversion.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Presented By : Murad Tukan
Robust estimation Problem: we want to determine the displacement (u,v) between pairs of images. We are given 100 points with a correlation score computed.
Real-Time Face Detection and Tracking Using Multiple Cameras RIT Computer Engineering Senior Design Project John RuppertJustin HnatowJared Holsopple This.
Tal Mor  Create an automatic system that given an image of a room and a color, will color the room walls  Maintaining the original texture.
Wavelet-Based Multiresolution Matching for Content-Based Image Retrieval Presented by Tienwei Tsai Department of Computer Science and Engineering Tatung.
Computer Science 631 Lecture 7: Colorspace, local operations
Chapter 9.  Mathematical morphology: ◦ A useful tool for extracting image components in the representation of region shape.  Boundaries, skeletons,
Table 3:Yale Result Table 2:ORL Result Introduction System Architecture The Approach and Experimental Results A Face Processing System Based on Committee.
CS 6825: Binary Image Processing – binary blob metrics
COLOR HISTOGRAM AND DISCRETE COSINE TRANSFORM FOR COLOR IMAGE RETRIEVAL Presented by 2006/8.
Digital Image Processing CCS331 Relationships of Pixel 1.
1 Webcam Mouse Using Face and Eye Tracking in Various Illumination Environments Yuan-Pin Lin et al. Proceedings of the 2005 IEEE Y.S. Lee.
CS654: Digital Image Analysis Lecture 25: Hough Transform Slide credits: Guillermo Sapiro, Mubarak Shah, Derek Hoiem.
1 Artificial Intelligence: Vision Stages of analysis Low level vision Surfaces and distance Object Matching.
Copyright Howie Choset, Renata Melamud, Al Costa, Vincent Lee-Shue, Sean Piper, Ryan de Jonckheere. All Rights Reserved Computer Vision.
Eye regions localization Balázs Harangi – University of Debrecen Ciprian Pop – Technical University of Cluj-Napoca László Kovács – University of Debrecen.
Introduction to Computer Graphics
Autonomous Robots Vision © Manfred Huber 2014.
Face Image-Based Gender Recognition Using Complex-Valued Neural Network Instructor :Dr. Dong-Chul Kim Indrani Gorripati.
A generic face processing framework: Technologies, Analyses and Applications Supervised by: Prof. Michael R. Lyu Presented by Jang Kim Fung Oral Defense.
October 16, 2014Computer Vision Lecture 12: Image Segmentation II 1 Hough Transform The Hough transform is a very general technique for feature detection.
An Image Retrieval Approach Based on Dominant Wavelet Features Presented by Te-Wei Chiang 2006/4/1.
Chapter 1: Image processing and computer vision Introduction
Image Segmentation Image segmentation (segmentace obrazu)
Student Gesture Recognition System in Classroom 2.0 Chiung-Yao Fang, Min-Han Kuo, Greg-C Lee, and Sei-Wang Chen Department of Computer Science and Information.
Color Image Processing
Color Image Processing
FINGER PRINT RECOGNITION USING MINUTIAE EXTRACTION FOR BANK LOCKER SECURITY Presented by J.VENKATA SUMAN ECE DEPARTMENT GMRIT, RAJAM.
Color Image Processing
Introduction to Computational and Biological Vision Keren shemesh
A language assistant system for smart glasses
Fast Preprocessing for Robust Face Sketch Synthesis
Computer Vision Lecture 4: Color
Fitting Curve Models to Edges
Chapter 1: Image processing and computer vision Introduction
Color Image Processing
Color Image Processing
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
Image segmentation Grey scale image Binary image
Introduction to Artificial Intelligence Lecture 22: Computer Vision II
Presentation transcript:

LYU0203 Smart Traveller with Visual Translator for OCR and Face Recognition Supervised by Prof. LYU, Rung Tsong Michael Prepared by: Wong Chi Hang Tsang Siu Fung Department of Computer Science & Engineering The Chinese University of Hong Kong

Outline Introduction Overall Design Korean OCR Face Detection Future Work

Introduction – What is VTT? Smart Traveller with Visual Translator (VTT)  Mobile Device which is convenient for a traveller to carry Mobile Phone, Pocket PC, Palm, etc.  Recognize and translate the foreign text into native language  Detect and recognize the face into name

Introduction – Motivation More and more people have mobile device which include Pocket PC, Palm, mobile phone. Mobile Device becomes more powerful. There are many people travelling aboard

Introduction – Motivation (Cont.) Types of programs for Mobile Device  Communication and Network  Multimedia  Games  Personal management  System tool  Utility

Introduction – Motivation (Cont.) Application for traveller?  Almost no!!! Very often, travellers encounter many problems about unfamiliar foreign language Therefore, the demand of an application for traveller is very large.

Introduction – Objective Help travellers to overcome language and memory power problems Two main features:  Recognize and translate Korean to English (Korean is not understandable for us)  Detect and recognize the face (Sometimes we forget the name of a friend)

Introduction – Objective (Cont.) Target of Korean OCR  Signs and Guideposts Printed Characters Contrast Text Color and Background Color Target of Face Recognizer  One face in photo  Frontal face  Limited set of faces

Introduction – Objective (Cont.) Real Life Examples  Sometimes we lose the way, we need to know where we are.  Sometimes we forget somebody we met before.

Overall Design of VTT System GUI Camera API Camera Korean OCRFace Recognizer Face Database Stroke Database & Dictionary Request Data RequestOutput User QueryResultQueryUpdateResult Request Response

KOCR – Design

KOCR – Text Area Detection Edge Detection using Sobel Filter

KOCR – Text Area Detection (Cont.) Horizontal and Vertical Edge Projection Hor izon tal Proj ecti on Threshold Vertical Projection

KOCR – Binarization Color Segmentation  Base on Color Histogram Threshold

KOCR – Stroke Extraction Labeling of Connected Component with 8- connectivity

KOCR – Stroke Extraction (Cont.) Why do we choose stroke but not whole character?  Korean Character is composed of Some Stroke types  Limited Set of Stroke Types in Korean

KOCR – Stroke Feature Our Proposed Feature  Five rays each side  Difference of adjacent rays (-1 or 0 or 1)  Has holes (0 or 1)  Dimension ratio of Stroke (width/height) (-1 or 0 or 1)

KOCR – Stroke Feature (Cont.) Problems Faced  Train the stroke database needs much time  Two or more strokes maybe stick together

KOCR – Stroke Recognition Exact Matching by Pre-learned Stroke Features Trained Decision Tree

KOCR – Pattern Identification Six Pattern of Korean Character Identify by simple if-then-else statement

Face Detection Outline 1. Find Face Region 2. Find the potential eye region 3. Locate the iris 4. Improvement

1. Find Face Region There are three methods available 1. Projection of the image 2. Base on gray-scale image 3. Color-based model

1. Find Face Region -Projection of the image Consider only one single color: blue, green or red. Usually blue pixel value is used because it can avoid the interference of the facial feature. Project the blue pixel vertically to find the left and right edge of face.

1. Find Face Region (Cont.) -Projection of the image x Edge of face Sum of pixel value

1. Find Face Region ( Cont. ) -Projection of the image The image should be filtered out the high frequency of this curve by FTT (Fast Fourier Transform) Assume the face occupy large area of the image

1. Find Face Region -Base on gray-scale image No color information Pattern recognition

1. Find Face Region -Color-based model We use this method because of its simplicity and robustness. Color-based model is used to represent color. Since human retina has three types of color photoreceptor cone cell, color model need three numerical components.

Color-based model ( Cont. ) There are many color model such as RGB, YUV (luminance-chrominance) and HSB (hue, saturation and brightness) Usually RGB color model will be transformed to other color model such as YUV and HSB.

Color-based model ( Cont. ) -YUV We use YUV or YC b C r color model. Y component is used to represent the intensity of the image C b and C r are used to represent the blue and red component respectively.

Color-based model ( Cont. ) -YC b C r Image Y Cb Cr Original Image -

Representation of Face color How can YUV color model represent face color? What happens when we transform the pixel into Cr-Cb histogram?

Representation of Face color We just use a simple ellipse equation to model skin color. Cb Cr

Representation of Face color where L is the length of the long axis and S is the length of the short axis. We choose L = 35.42, S = , θ = (radius) The equation of the ellipse :

Representation of Face color - Color segmentation The white regions represent the skin color pixels

Representation of Face color -Color segmentation (modified version1) We distribute some agents in the image uniformly. Then each agent will check whether the pixel is a skin-like pixel and not visited by the other agent. If yes, it will produce 4 more agents at its four neighboring points. If no, it will moved to one of its four neighboring points randomly.

Representation of Face color ( Cont. ) -Color segmentation (modified version1) This agent produce 4 more agents If the pixel is a skin-like pixel and not visited by the other agent, produce 4 more agents at its four neighboring points

Representation of Face color ( Cont. ) -Color segmentation (modified version1) Otherwise, it will moved to one of its four neighboring points randomly This agent move to one of four neighboring point

Representation of Face color ( Cont. ) -Color segmentation (modified version1) Each agent will search their own region Each region are shown in the next slide with different color.

Representation of Face color ( Cont. ) -Color segmentation (modified version1) The advantage of this algorithm is that we need not to search the whole image. Therefore, it is fast.

Representation of Face color ( Cont. ) -Color segmentation (modified version1) of pixels is searched (about 18.7%) There are 37 regions

2. Eye detection After the segmentation of face region, we have some parts which are not regarded as skin color. They are probably the region of eye and mouth We only consider the red component of these regions because it usually includes the most information about faces.

2. Eye detection ( Cont. ) We extraction such regions by pseudo- convex hull.

2. Eye detection ( Cont. ) We do the following on the regions of potential eye region 1. Histogram equalization 2. Threshold

2. Eye detection ( Cont. ) Histogram equalization Threshold with < 49 After the histogram equalization and threshold, the searching space of eyes is greatly reduced.

3. Locate the iris After the operations above, we almost find the eye. However, we should locate the iris. We use the following different methods  Template matching  Hough Transform

3. Locate the iris ( Cont. ) -Template matching It bases on normalized cross-correlation. It is used to measure the similarity between two images

3. Locate the iris ( Cont. ) -Template matching Let I 1, I 2 be images of the same size. I 1 (p i ) = a i, I 2 (p i ) = b i NCC(I 1, I 2 ) lies on the range [-1, 1]

3. Locate the iris ( Cont. ) -Template matching We use this template and calculate the NCC. This template can be obtained by averaging all the eye image.

3. Locate the iris ( Cont. ) -Template matching Red region show the result

3. Locate the iris ( Cont. ) -Hough transform Hough Transform can find the complete shape of the edge according to small portion of edge information. It works with a parametric representation of the object we are looking for. We use Hough transform with 2D circle parametric representation to find the iris.

3. Locate the iris ( Cont. ) -Hough transform We find the edge of eye by Sobel filter.

3. Locate the iris ( Cont. ) -Hough transform We apply a circle on the edge image and count the number of pixel lying on the circle

3. Locate the iris ( Cont. ) -Hough transform A(x,y,r) <- Number of pixel where A(x,y,r) is Accumulator, where x,y are the coordinate of the center and r is the radius of the circle. The searching space for the circle is [x, y, r] = [17, 17, 8].

3. Locate the iris ( Cont. ) -Hough transform We have tried this method It fails to find the iris

4. Improvement Skin Color Detection  Neuron Network with simplified activate function (polynomial)  Probability function (e.g. Bayesian estimation) Setup face Shape model it estimates the shape of face

4. Improvement ( Cont. ) Template Matching Replace it with deformable template or probability function.

Future Work Stroke Combination Dictionary Face Detection Improvement Face Recognition  normal luminance light source  about 20 people  > 90 % accuracy Port the system into Pocket PC

Q&A

~The End~