Syntactical Pattern Recognition with 2-Dimensional Array Grammars and Array Automata Faculty of Informatics Vienna University of Technology Wien, Austria.

Slides:



Advertisements
Similar presentations
Chapter 5 Pushdown Automata
Advertisements

Tiling Automata: a computational model for recognizable two-dimensional languages Marcella Anselmo Dora Giammarresi Maria Madonia Univ. of Salerno Univ.
Chap. 5, Top-Down Parsing J. H. Wang Mar. 29, 2011.
CFGs and PDAs Sipser 2 (pages ). Long long ago…
作者 : 陳鍾誠 單位 : 金門技術學院資管系 URL : 日期 : 2015/5/5 剖析器 - Parser.
ICE1341 Programming Languages Spring 2005 Lecture #4 Lecture #4 In-Young Ko iko.AT. icu.ac.kr iko.AT. icu.ac.kr Information and Communications University.
Handwritten Character Recognition Using Artificial Neural Networks Shimie Atkins & Daniel Marco Supervisor: Johanan Erez Technion - Israel Institute of.
Applied Computer Science II Chapter 2 : Context-free languages Prof. Dr. Luc De Raedt Institut für Informatik Albert-Ludwigs Universität Freiburg Germany.
Functional Design and Programming Lecture 9: Lexical analysis and parsing.
Top-Down Parsing.
Foundations of (Theoretical) Computer Science Chapter 2 Lecture Notes (Section 2.1: Context-Free Grammars) David Martin With some.
Parsing — Part II (Ambiguity, Top-down parsing, Left-recursion Removal)
Normal forms for Context-Free Grammars
Algorithmic Problems in Algebraic Structures Undecidability Paul Bell Supervisor: Dr. Igor Potapov Department of Computer Science
Table-driven parsing Parsing performed by a finite state machine. Parsing algorithm is language-independent. FSM driven by table (s) generated automatically.
Cs466(Prasad)L8Norm1 Normal Forms Chomsky Normal Form Griebach Normal Form.
MIT Top-Down Parsing Martin Rinard Laboratory for Computer Science Massachusetts Institute of Technology.
Grammars, Languages and Finite-state automata Languages are described by grammars We need an algorithm that takes as input grammar sentence And gives a.
E.G.M. PetrakisBinary Image Processing1 Binary Image Analysis Segmentation produces homogenous regions –each region has uniform gray-level –each region.
Handwritten Character Recognition using Hidden Markov Models Quantifying the marginal benefit of exploiting correlations between adjacent characters and.
1 Two-dimensional Context-Free Grammars: Mathematical Formulae Recognition Daniel Průša, Václav Hlaváč Center for Machine Perception Faculty of Electrical.
FEATURE EXTRACTION FOR JAVA CHARACTER RECOGNITION Rudy Adipranata, Liliana, Meiliana Indrawijaya, Gregorius Satia Budhi Informatics Department, Petra Christian.
1 Mathematical Institute Serbian Academy of Sciences and Arts, Belgrade DEUKS Meeting Valencia, September 9-11, 2008, Valencia New PhD modules proposal.
Scattered Context Grammars Alexander Meduna Faculty of Information Technology Brno University of Technology Brno, Czech Republic, Europe.
DIPLOMA THESIS Peter Černo Clearing Restarting Automata Supervised by RNDr. František Mráz, CSc.
Chapter 9 Syntax Analysis Winter 2007 SEG2101 Chapter 9.
APPLICATIONS OF CONTEXT FREE GRAMMARS BY, BRAMARA MANJEERA THOGARCHETI.
Review: –How do we define a grammar (what are the components in a grammar)? –What is a context free grammar? –What is the language defined by a grammar?
Top-Down Parsing - recursive descent - predictive parsing
Chapter 10: Compilers and Language Translation Invitation to Computer Science, Java Version, Third Edition.
Probabilistic Context Free Grammars for Representing Action Song Mao November 14, 2000.
Languages A Language L over a finite alphabet  is a set of strings of characters from the alphabet. Examples are : 1.L = { s | s is a grammatical English.
Some Probability Theory and Computational models A short overview.
COMPARISON OF IMAGE ANALYSIS FOR THAI HANDWRITTEN CHARACTER RECOGNITION Olarik Surinta, chatklaw Jareanpon Department of Management Information System.
State Complexity: Recent Results and Future Directions Sheng Yu Department of Computer Science University of Western Ontario London, Ontario,
GRAMMATICAL INFERENCE OF LAMBDA-CONFLUENT CONTEXT REWRITING SYSTEMS Peter Černo Department of Computer Science Charles University in Prague, Faculty of.
1 Computability Five lectures. Slides available from my web page There is some formality, but it is gentle,
COP 4620 / 5625 Programming Language Translation / Compiler Writing Fall 2003 Lecture 3, 09/11/2003 Prof. Roy Levow.
Online Kinect Handwritten Digit Recognition Based on Dynamic Time Warping and Support Vector Machine Journal of Information & Computational Science, 2015.
Parsing Introduction Syntactic Analysis I. Parsing Introduction 2 The Role of the Parser The Syntactic Analyzer, or Parser, is the heart of the front.
Center for PersonKommunikation P.1 Background for NLP Questions brought up by N. Chomsky in the 1950’ies: –Can a natural language like English be described.
Overview of Previous Lesson(s) Over View  Symbol tables are data structures that are used by compilers to hold information about source-program constructs.
1 / 48 Formal a Language Theory and Describing Semantics Principles of Programming Languages 4.
CSE 425: Syntax I Syntax and Semantics Syntax gives the structure of statements in a language –Allowed ordering, nesting, repetition, omission of symbols.
CSC312 Automata Theory Lecture # 26 Chapter # 12 by Cohen Context Free Grammars.
Concepts and Realization of a Diagram Editor Generator Based on Hypergraph Transformation Author: Mark Minas Presenter: Song Gu.
1 CD5560 FABER Formal Languages, Automata and Models of Computation Lecture 3 Mälardalen University 2007.
LECTURE 4 Syntax. SPECIFYING SYNTAX Programming languages must be very well defined – there’s no room for ambiguity. Language designers must use formal.
How YACC is constructed. How Yacc works To construct a parsing machine for arithmetic expressions, a special case considered to simplify the account of.
Chapter 2 Scanning. Dr.Manal AbdulazizCS463 Ch22 The Scanning Process Lexical analysis or scanning has the task of reading the source program as a file.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 3 Sept 16, 2005 Nanjing University of Science & Technology.
1 Course Overview Why this course “formal languages and automata theory?” What do computers really do? What are the practical benefits/application of formal.
COMP 3438 – Part II-Lecture 6 Syntax Analysis III Dr. Zili Shao Department of Computing The Hong Kong Polytechnic Univ.
Refining the Nonterminal Complexity of Graph-controlled Grammars Faculty of Informatics TU Wien, Wien, Austria Rudolf FREUND Wilhelm-Schickard-Institut.
1 CD5560 FABER Formal Languages, Automata and Models of Computation Lecture 3 Mälardalen University 2006.
Optical Character Recognition
Modeling Arithmetic, Computation, and Languages Mathematical Structures for Computer Science Chapter 8 Copyright © 2006 W.H. Freeman & Co.MSCS SlidesAlgebraic.
Finite Automata.
Programming Languages Translator
José Manuel Iñesta José Martínez Sotoca Mateo Buendía
The chomsky hierarchy Module 03.3 COP4020 – Programming Language Concepts Dr. Manuel E. Bermudez.
PARSE TREES.
Subject Name:COMPILER DESIGN Subject Code:10CS63
Lexical and Syntax Analysis
Jaya Krishna, M.Tech, Assistant Professor
Compilers Principles, Techniques, & Tools Taught by Jing Zhang
SMEM Algorithm for Mixture Models
Automaton-like P Colonies: New Variants
Compilers Principles, Techniques, & Tools Taught by Jing Zhang
Presentation transcript:

Syntactical Pattern Recognition with 2-Dimensional Array Grammars and Array Automata Faculty of Informatics Vienna University of Technology Wien, Austria Rudi FREUND

Offline Character Recognition Overview ● Preprocessing - normalization - elimination of noisy pixels - thinning ● Syntactic Analysis Variants of Array Grammars/Automata for the Syntactic Analysis Summary Related Theoretical Results

Character Recognition – Preprocessing Normalization The scanned characters first are normalized to fill out a 320×400 grid in order to get comparable patterns. Then noisy pixels are eliminated. After noise elimination, the resulting arrays on the 320×400 grid are mapped on a 20×25 grid.

Character Recognition – Preprocessing Elimination of Noisy Pixels Algorithms are based on parallel array grammars eliminate pixel if number of pixels in the 8- neighbourhood = 0, iterate this algorithm until no more changes

Character Recognition – Preprocessing Thinning Algorithms are based on parallel array grammars. Reduction to a unitary skeleton (e.g., J. H. Sossa, An improved parallel algorithm for thinning digital patterns, Pattern Recognition Letters 10 (1989), pp ).

Character Recognition – Preprocessing Result Unitary skeleton on a 20 x 25 grid

Variants of Array Grammars/Automata for the Syntactic Analysis (SSPR’96) Henning Fernau, Rudolf Freund: Bounded parallelism in array grammars used for character recognition. In: Petra Perner, Patrick Shen-Pei Wang, Azriel Rosenfeld (Eds.): Advances in Structural and Syntactical Pattern Recognition, 6th International Workshop, SSPR'96, Leipzig, Germany, August 20-23, 1996, Proceedings. LNCS 1121, Springer, Berlin, 1996,

Bounded Parallelism

Bounded Parallelism/Prescribed Teams array productions in team are applied in parallel

Bounded Parallelism/Prescribed Teams derivation modes array productions in team are applied in parallel; the teams themselves may be applied in different derivation modes (variants of co-operation as in co-operating distributed grammar systems): =k, >k, <k, *, t (maximal derivation mode) internally hybrid modes: (t,=k), (t,>k), (t, m,<k)

Bounded Parallelism/Prescribed Teams finite index restriction/pattern analysis Finite index restriction: by applying the array productions of a team in parallel, all non-terminal symbols in the current sentential form must be affected Analysis of a given pattern: whenever a terminal symbol is generated, it must coincide with the symbol in the pattern (character) to be analysed

Syntactic Pattern Analysis – Distance from Ideal Cluster ● deviation of lines ● gaps in lines ● superfluous/missing lines ● superfluous (remaining) pixels During the analysis of a given pattern, its distance from the ideal cluster of arrays representing a specific character is computed. Features to be taken into account:

Variants of Array Grammars/Automata for the Syntactic Analysis (SSPR’98, k-head finite automata) Henning Fernau, Rudolf Freund, Markus Holzer: Character recognition with k-head finite array automata. In: Adnan Amin, Dov Dori, Pavel Pudil, Herbert Freeman (Eds.): Advances in Pattern Recognition, Joint IAPR International Workshops SSPR '98 and SPR '98, Sydney, NSW, Australia, August 11-13, 1998, Proceedings. LNCS 1451, Springer, Berlin, 1998,

k-head finite automata The k heads in a k-head finite array automaton are the counterparts of the k non-terminal symbols in array grammars with prescribed teams and finite index. In each step, every head has to move. The automaton has a „head sensing ability“, i.e., two heads can never occupy the same position. Moreover, a position carrying a terminal symbol in the array to be parsed can only be visited once by one of the k heads (and then is marked as „forbidden position“ for the rest of the parsing procedure).

Variants of Array Grammars/Automata for the Syntactic Analysis (regulated array grammars of finite index) Henning Fernau, Rudolf Freund, Markus Holzer: Regulated array grammars of finite index. II: Syntactic pattern recognition. In: Gheorghe Păun, Arto Salomaa (Eds.): Grammatical Models of Multi-Agent Systems. Topics in Computer Mathematics 8, Gordon and Breach Science Publishers, 1999,

Variants of Array Grammars/Automata for the Syntactic Analysis (SSPR 2000) Rudolf Freund, Markus Neubauer, Martin Summerer, Stefan Gruber, Jürgen Schaffer, Roland Swoboda: A hybrid system for the recognition of hand-written characters. In: Francesc J. Ferri, José Manuel Iñesta Quereda, Adnan Amin, Pavel Pudil (Eds.): Advances in Pattern Recognition, Joint IAPR International Workshops SSPR 2000 and SPR 2000, Alicante, Spain, August 30 - September 1, 2000, Proceedings. LNCS 1876, Springer, Berlin, 2000,

Hybrid Systems Pre-selection by neural network (for a given pattern, only a few pre-selected array grammars have to analyse it) Application of teams controlled by additional mechanism (programmed, matrix) Look-ahead instead of inefficient backtracking (larger neighbourhoods)

Related Theoretical Results Henning Fernau, Rudolf Freund, Markus Holzer: Regulated array grammars of finite index. I: Theoretical Investigations. In: Gheorghe Păun, Arto Salomaa (Eds.): Grammatical Models of Multi-Agent Systems. Topics in Computer Mathematics 8, Gordon and Breach Science Publishers, 1999,

Regulated Array Grammars of Finite Index Restricted by the finite index condition, with the control mechanisms of using variants of context- fee array productions in prescibed teams, matrices or a control graph, the corresponding families of generated array languages coincide, even in the appearance checking case. The corresponding models of k- head automata accept the same families of array languages, too. With respect to k, we obtain infinite hierarchies for dimensions n > 1.

E N D

THANK YOU FOR YOUR ATTENTION GRAZIE, MERCI, KIITOS, DĺKY, DANKE!