Download presentation
Presentation is loading. Please wait.
1
Lexical and Syntax Analysis
Chapter 4 Lexical and Syntax Analysis
2
Chapter 4 Topics Introduction Lexical Analysis The Parsing Problem
Recursive-Descent Parsing Bottom-Up Parsing Copyright © 2006 Addison-Wesley. All rights reserved.
3
Why study lexical and syntax analyzers?
Shows application of grammars discussed in chapter 3 Lexical and syntax analysis not just used in compiler design: program listing formatters programs that compute complexity programs that analyze and react to configuration files Good to have some background, since compiler design not required Copyright © 2006 Addison-Wesley. All rights reserved.
4
Unsolicited email from student…
Hiyo, For my internship (and hopefully soon full-time job) one of the projects I'm doing is managing all of the config files for all the servers and devices we have. They're generated from templates, and then manual updates are also added, and we want to keep track of all the data, and make sure all of those configs are up to date and correct, mirror what we think it should be in the database, and a few other things to end a lot of late-night headaches. A lot of the aspects of this have me going over notes again from Programming Languages, in dealing with understanding syntax, parsing, etc. Thought you might enjoy seeing those lessons being applied in the "real world"! Copyright © 2006 Addison-Wesley. All rights reserved.
5
Introduction Language implementation systems must analyze source code, regardless of the specific implementation approach (compiled, interpreted, hybrid) Nearly all syntax analysis is based on a formal description of the syntax of the source language (context-free grammars or BNF) The parser can be based directly on the BNF Parsers based on BNF are easy to maintain (modular) Copyright © 2006 Addison-Wesley. All rights reserved.
6
Syntax Analysis The syntax analysis portion of a language processor nearly always consists of two parts: A low-level part called a lexical analyzer (mathematically, a finite automaton based on a regular grammar). Also called a scanner. A high-level part called a syntax analyzer, or parser (mathematically, a push-down automaton based on a context-free grammar, or BNF) Copyright © 2006 Addison-Wesley. All rights reserved.
7
Reasons to Separate Lexical and Syntax Analysis
Simplicity - less complex approaches can be used for lexical analysis; separating them simplifies the parser Efficiency - separation allows optimization of the lexical analyzer – lex is fast! Portability - parts of the lexical analyzer may not be portable, but the parser always is portable Why? Copyright © 2006 Addison-Wesley. All rights reserved.
8
Lexical Analysis A lexical analyzer is a pattern matcher for character strings A lexical analyzer is a “front-end” for the parser. Views source as single string of characters. Identifies substrings of the source program that belong together - lexemes Lexemes match a character pattern, which is associated with a lexical category called a token, normally represented as an enum/int sum is a lexeme; its token may be IDENT Copyright © 2006 Addison-Wesley. All rights reserved.
9
Lexical Analysis (review)
result = oldsum – value / 100; Lexeme Token result IDENT = ASSIGN_OP oldsum - SUBTRACT_OP value / DIVISION_OP 100 INT_LIT ; SEMICOLON Copyright © 2006 Addison-Wesley. All rights reserved.
10
Lexical Analysis (continued)
The lexical analyzer is usually a function that is called by the parser when it needs the next token (used to process entire file at one time) Skips comments and whitespace May insert lexemes for user-defined names into symbol table Detects syntax errors in tokens, such as ill-formed floating point literals oldsum ident - subtract_op = assign_op value ident result ident … Scanner result = oldsum – value / 100; Copyright © 2006 Addison-Wesley. All rights reserved.
11
Lexical Analysis (continued)
Three approaches to building a lexical analyzer: Write a formal description of the tokens based on regular expressions and use a software tool that constructs a lexical analyzer given such a description (e.g., lex, flex) Design a state diagram that describes the tokens and write a program that implements the state diagram Design a state diagram that describes the tokens and hand-construct a table-driven implementation of the state diagram Copyright © 2006 Addison-Wesley. All rights reserved.
12
State Diagram to recognize names, reserved words and integer literals – regular languages
A naïve state diagram would have a transition from every state on every character in the source language - such a diagram would be very large! Use character class LETTER for all 52 letters, allows single transition from Start Instead of using state diagram for reserved words, use lookup table Use DIGIT to simplify transition for numbers Note that after 1st character, LETTER or DIGIT is allowed in names Copyright © 2006 Addison-Wesley. All rights reserved.
13
Lexical Analysis (cont.)
Implementation (assume initialization): int lex() { getChar(); // puts char in nextChar, type in charClass // skips white space switch (charClass) { case LETTER: // Notice correspondence to state transition addChar(); // appends nextChar onto lexeme getChar(); while (charClass == LETTER || charClass == DIGIT) { addChar(); } return lookup(lexeme); break; … Copyright © 2006 Addison-Wesley. All rights reserved.
14
Lexical Analysis (cont.)
… case DIGIT: addChar(); getChar(); while (charClass == DIGIT) { } return INT_LIT; break; } /* End of switch */ } /* End of function lex */ so this corresponds to… How would 123abc be interpreted? Copyright © 2006 Addison-Wesley. All rights reserved.
15
Compiler Tools Now is a good time to look at lex/flex
Copyright © 2006 Addison-Wesley. All rights reserved.
16
The Parsing Problem (syntax analysis)
Goals of the parser, given an input program: Find all syntax errors; for each, produce an appropriate diagnostic message, and recover quickly (to check rest of program) Produce the parse tree, or at least a trace of the parse tree, for the program. The parse tree is used as the basis for translation. Parse tree contains all the information needed by a language processor Copyright © 2006 Addison-Wesley. All rights reserved.
17
The Parsing Problem (cont.)
Two categories of parsers Top down - produce the parse tree, beginning at the root Order is that of a leftmost derivation Traces or builds the parse tree in preorder Bottom up - produce the parse tree, beginning at the leaves Order is that of the reverse of a rightmost derivation Parsers look only one token ahead in the input Copyright © 2006 Addison-Wesley. All rights reserved.
18
Which will we study? Both!
First we’ll look at top-down techniques (recursive descent parsing) and problems with left-recursive grammars. We’ll code a basic recursive descent parser. Then we’ll look at bottom-up parsing and how to use yacc/bison Copyright © 2006 Addison-Wesley. All rights reserved.
19
The Parsing Problem: Top-Down
Non-terminal string of terminals mixed string Consider the sentential form xA Top-down Parsers parser must choose the correct rule for nonterminal A to get the next sentential form in the leftmost derivation, using only the first token produced by A Notational conventions: Terminal symbols – lowercase letters at beginning of alphabet (a, b, ...) Nonterminal symbols – uppercase letters at beginning of alphabet (A, B, ...) Terminals or nonterminals – uppercase letters at end of alphabet (W,X,Y,Z) Strings of terminals – lowercase letters at end of alphabet (w,x,y,z) Mixed strings (terminals and/or nonterminals) – lowercase Greek letters () Example: a = <term> + <expr> Copyright © 2006 Addison-Wesley. All rights reserved.
20
Top-Down Parsing, continued
Parsing xA Assume three A-rules are: A->bB | cBb | a Next sentential form could be xbB, xcBb or xa Compare next token of input with first symbol generated by each rule. Not always easy, could be a nonterminal. May need to backtrack. What does that mean? Copyright © 2006 Addison-Wesley. All rights reserved.
21
Backtrack Example Parsing ba Assume the rules are (start symbol is A):
A->bE |Ba B -> b E -> e Copyright © 2006 Addison-Wesley. All rights reserved.
22
The Complexity of Parsing
Parsers that work for any unambiguous grammar are complex and inefficient ( O(n3), where n is the length of the input ). Algorithms must frequently back up and reparse, requires more maintenance of tree. Too slow to be used in practice. Compilers use parsers that only work for a subset of all unambiguous grammars, but do it in linear time ( O(n), where n is the length of the input ) Copyright © 2006 Addison-Wesley. All rights reserved.
23
Types of Top-Down Parsers
The most common top-down parsing algorithms: Recursive descent - a coded implementation LL parsers - table driven implementation L left-to-right scan, L leftmost derivation (LL) ANTLR is an LL(*) recursive-descent parser generator LL(*) means not restricted to a finite k tokens of lookahead Copyright © 2006 Addison-Wesley. All rights reserved.
24
Recursive-Descent Parsing
Hand-coded solution – general approach Write a subprogram for each nonterminal (e.g., <expr>) Subprograms may be recursive (e.g., for nested structures) Collect all subprograms Produces a parse tree in top-down order EBNF is ideally suited to be the basis for a recursive-descent parser, because EBNF minimizes the number of nonterminals (i.e., fewer subprograms) Copyright © 2006 Addison-Wesley. All rights reserved.
25
Recursive-Descent Parsing (cont.)
A grammar for simple expressions: <expr> <term> {(+ | -) <term>} <term> <factor> {(* | /) <factor>} <factor> id | ( <expr> ) A grammar for selection and declaration <if-statement> if <logic_expr><statement>[else <statement>] <ident-list> ident {, <ident>} Grammar does not enforce associativity rules, compiler must Copyright © 2006 Addison-Wesley. All rights reserved.
26
Recursive-Descent Parsing (cont.)
Assume we have a lexical analyzer function named lex, which puts the next token code in nextToken Simple case: only one RHS for a non- terminal For each terminal symbol in the RHS (e.g., if), compare it with the next input token; if they match, continue, else there is an error For each nonterminal symbol in the RHS (e.g., <term>), call its associated parsing subprogram Example to follow… Copyright © 2006 Addison-Wesley. All rights reserved.
27
Recursive-Descent Parsing (cont.)
/* Function expr Parses strings in the language generated by the rule: <expr> → <term> {(+ | -) <term>} */ void expr() { // Parse the first term term(); /* As long as the next token is + or -, call lex to get the next token, and parse the next term */ while (nextToken == PLUS_CODE || nextToken == MINUS_CODE){ lex(); // get next token term(); } } This particular routine does not detect errors Convention: Every parsing routine leaves the next token in nextToken Copyright © 2006 Addison-Wesley. All rights reserved.
28
Recursive-Descent Parsing (cont.)
More complex: Nonterminal has more than one RHS. Requires an initial process to determine which RHS it is to parse. The correct RHS is chosen on the basis of the next token of input (the lookahead) The next token is compared with the first token that can be generated by each RHS until a match is found If no match is found, it is a syntax error Copyright © 2006 Addison-Wesley. All rights reserved.
29
Recursive-Descent Parsing (cont.)
/* Function factor Parses strings in the language generated by the rule: <factor> -> id | ( <expr> ) */ void factor() { // Determine which RHS if (nextToken) == ID_CODE) lex(); // For the RHS id, just call lex /* If the RHS is ( <expr> ) – call lex to pass over the left parenthesis, call expr, then check for the right parenthesis */ else if (nextToken == LEFT_PAREN_CODE) { lex(); // skip over ( expr(); // ensure we have an expression if (nextToken == RIGHT_PAREN_CODE) lex(); // ensure the next is a ) else error(); } /* End of else if (nextToken == ... */ else error(); /* Neither RHS matches */ } Copyright © 2006 Addison-Wesley. All rights reserved.
30
Recursive-Descent Parsing (cont.)
/* Function term Parses strings in the language generated by the rule: <term> -> <factor> {(* | /) <factor>} */ void term() { /* Parse the first factor */ factor(); /* Keep processing term as long as * or / */ while (nextToken == AST_CODE || nextToken == SLASH_CODE) { lex(); // just skip over * and / factor(); } Copyright © 2006 Addison-Wesley. All rights reserved.
31
Recursive-Descent Parsing (cont.)
Trace the code for a + b Call lex // sets nextToken to a, ID_CODE Enter <expr> Enter <term> Enter <factor> // a is factor Call lex // sets nextToken to +, PLUS_CODE Exit <factor> Exit <term> // only 1 factor, no * or / + matches PLUS_CODE // while loop Call lex // sets nextToken to b, ID_CODE Enter <factor> // b is factor Call lex // sets nextToken to end-of-input Exit <factor> Exit <term> Exit <expr> //end while Copyright © 2006 Addison-Wesley. All rights reserved.
32
Recursive-Descent Parsing (cont.)
Parse tree for a + b <expr> <term> + + <term> <factor> <factor> A B Copyright © 2006 Addison-Wesley. All rights reserved.
33
Recursive Descent Quick Exercise
Do Recursive Descent exercises 1 Copyright © 2006 Addison-Wesley. All rights reserved.
34
Recursive-Descent Parsing / LL Grammars and Left Recursion
Important to consider limitations of recursive- descent in terms of grammar restrictions The LL Grammar Class The Left Recursion Problem If a grammar has left recursion, either direct or indirect, it cannot be the basis for a top-down parser A grammar can be modified to remove left recursion A -> A + B // direct left recursive continues to try to parse A Left-to-right scan Leftmost Copyright © 2006 Addison-Wesley. All rights reserved.
35
Left Recursion Example
E -> E + T | T T -> T * F | F F -> (E) | id x + y + z nextToken = x E // terminal symbol not matched E + T // terminal symbol not matched E + T + T // terminal symbol not matched E + T + T + T // terminal symbol not matched A person, of course, would choose T rather than E + T, but the function for <expr> is deterministic, would always choose E + T Copyright © 2006 Addison-Wesley. All rights reserved.
36
Left Recursion– How to remove?
A . . . A -> A | Left-recursive string A->Aa->Aaa -> Aaaa -> Baaa A -> A’ A’ -> A’ | Right-recursive string A->BA’->BaA’->BaaA’->BaaaA’ ->Baaa R A . . . Both strings are followed by series of s Copyright © 2006 Addison-Wesley. All rights reserved.
37
Convince yourself (quick exercise)
Grammar with Left recursion A -> Ax | B -> z Grammar without left recursion A -> A’ A’ -> xA’ | B -> z We could just do x for this simple grammar, but see next slide Which grammar (or both) can produce: zxx zxxx Copyright © 2006 Addison-Wesley. All rights reserved.
38
Slightly more complex example
Grammar with Left recursion A -> Ax | Ay | Az | B -> c Grammar without left recursion A -> A’ A’ -> xA’ | yA’ | zA’ | B -> c Grammar without left recursion or erasure rule, suggested by a student: A -> A’ A’ -> xA’ | yA’ | zA’ | x | y | z B -> c Are these equivalent? Can we parse c?
39
LL Grammars/Removing Left Recursion
To remove direct left recursion: For each nonterminal, A, Group the A-rules as: A->A1, | ...|Am|1|...|n where none of the s begin with A Replace the original A-rules with A -> 1A' | 2A' | nA' A' -> 1A' | ... mA' | where is the empty string recursive rules erasure rule NOTE: The effect of original A rules is to add as to end of string, the revised rules add as to the end of the string, then get rid of the A Copyright © 2006 Addison-Wesley. All rights reserved.
40
Recursive-Descent Parsing / LL Grammars (cont)
E -> E + T | T T -> T * F | F F -> (E) | id E-Rules 1 = + T and = T E -> T E' E' -> + T E' | T-Rules 1 = * F and = F T -> F T' T' -> * F T' | Revised Grammar E -> T E' E' -> + T E' | e T -> F T' T' -> * F T' | e F -> (E) | id Think of T as base case Remember: A -> 1A' Remember: A' -> aA' Copyright © 2006 Addison-Wesley. All rights reserved.
41
Recursive-Descent Parsing / LL Grammars (cont)
Grammar generates the same language but is not left recursive. Ex: A + B * C E -> E + T | T T -> T * F | F F -> (E) | id E + T T + T F + T A + T A + T * F A + F * F A + B * F A + B * C T E' //rule 1 F T ' E ' // rule 3 A T’ E' // rule 5 F E ' // rule 4 erasure A + T E' // rule 2 A + F T' E’ // rule 3 A + B T' E’ // rule 5 A + B * F T' E’// rule 4 A + B * C T' E’// rule 5 A + B * C E’ // erasure rule A + B * C // erasure rule E -> T E' E' -> + T E' | T -> F T' T' -> * F T' | F -> (E) | id Copyright © 2006 Addison-Wesley. All rights reserved.
42
Convince Yourself E -> T E' E' -> + T E' | T -> F T' T' -> * F T' | F -> (E) | id E -> E + T | T T -> T * F | F F -> (E) | id Try A * C + B with both grammars (hint for 2nd grammar: try T + T) Copyright © 2006 Addison-Wesley. All rights reserved.
43
What about EBNF? Could be: E -> E + T | T E -> T { + T}
T -> F { * F} F -> (E) | id E -> E + T | T T -> T * F | F F -> (E) | id Translates to while loop in recursive descent
44
More complex example The general solution using BNF may be easier to deal with for more complex rules: E -> E + T | T | E ** F T -> T * F | F F -> (E) | id NOTE: The above grammar would not have correct precedence, the purpose is just to show that it might not always be so easy to convert left recursion to EBNF. When it can be done easily, converting to EBNF would be a good approach. E -> TE’ E’->+TE’ | **FE’ | T -> F T' T' -> * F T' | F -> (E) | id
45
Recursive-Descent Parsing / LL Grammars (cont)
Indirect left recursion poses same problem A -> B a A B -> A b A subprogram calls B, which calls A Algorithm to remove indirect left recursion not in text This is a problem for all top-down parsing algorithms, not just recursive descent Good news: When writing a grammar for a programming language, can usually avoid left recursion (both types) Copyright © 2006 Addison-Wesley. All rights reserved.
46
Left Recursion Exercise
Consider the simple grammar: CSM -> CSM a | ORE ORE -> g Show that this grammar can generate gaa. Rewrite the grammar to remove the left recursion. Show that it can still generate gaa. Copyright © 2006 Addison-Wesley. All rights reserved.
47
Recursive-Descent Parsing /LL Grammars & pairwise disjointness
Need to be able to choose RHS on the basis of the next token, using only first token generated by leftmost nonterminal Test to determine whether this can be done is pairwise disjointness test. Must compute a set named FIRST based on RHSs of nonterminal symbol Copyright © 2006 Addison-Wesley. All rights reserved.
48
Recursive-Descent Parsing /LL Grammars & pairwise disjointness
Pairwise Disjointness Test: For each nonterminal, A, in the grammar that has more than one RHS, for each pair of rules, A i and A j, it must be true that FIRST(i) FIRST(j) = Examples: A aB | bAb | Bb B -> cB | d FIRST of A-rules are {a}, {b} and {c, d} These are disjoint FIRST of B-rules are {c},{d} Also disjoint NOTE that we do the FIRST of all non-terminals, not just the first one Copyright © 2006 Addison-Wesley. All rights reserved.
49
Recursive-Descent Parsing /LL Grammars & pairwise disjointness
A -> aB | Bab B -> aB | g FIRST sets of A are {a} and {a, g} which are NOT disjoint If next input is an a, can't decide which RHS Becomes more complex as more RHSs begin with nonterminals Example: do top-down parse of ag using just a one-token lookahead. token is a. Do I apply aB or Bab? remember recursive descent coding: if (nextToken == ‘a’) // do what?? Copyright © 2006 Addison-Wesley. All rights reserved.
50
Pairwise Disjoint cautions
Simple Pairwise Disjoint A -> aB | aC // FIRST(A) = {a},{a} not disjoint B -> x // FIRST(B) = {x} C-> y // FIRST(C ) = {y} Is this pairwise disjoint? A -> aB | gC B -> x C -> x left Recursive is not the same as Pairwise disjoint A -> Aa | aB is this a problem? Copyright © 2006 Addison-Wesley. All rights reserved.
51
More examples Submitted by students in F08 A => aB | c
B => c | Ab <yum> => <fruit><dessert> | <veggie><dessert> <fruit> => apple | orange <veggie> => carrot | broccoli <dessert> => brownie | icecream <yum> => <fruit><dessert> | <fruit><veggie><dessert> <fruit> => apple | orange <veggie> => carrot | broccoli <dessert> => brownie | icecream Copyright © 2006 Addison-Wesley. All rights reserved.
52
Recursive-Descent Parsing (cont.)
Left factoring can sometimes resolve the problem Replace <variable> identifier | identifier [<expression>] with <variable> identifier <idarg> <idarg> | [<expression>] or <variable> identifier [[<expression>]] (the outer brackets are metasymbols of EBNF indicating that inside is optional) Copyright © 2006 Addison-Wesley. All rights reserved.
53
Pairwise Disjoint Exercise
Do the Pairwise Disjoint Exercise Copyright © 2006 Addison-Wesley. All rights reserved.
54
Bottom-Up Parsing – the Idea
Example grammar and derivation: S -> aAc A -> aA | b S => aAc => aaAc => aabc grammar sample derivation Starting with sentence, aabc, must find handle. String aabc contains RHS b. Yields aaAc. String aaAc contains handle aA. Yields aAc. String aAc contains handle aAc. Yields S. Copyright © 2006 Addison-Wesley. All rights reserved.
55
The Parsing Problem: Bottom-Up
Bottom-up parsers Given a right sentential form, , determine what substring of is the RHS of the rule in the grammar that must be reduced to produce the previous sentential form in the right derivation Given sentential form may include more than one RHS from the language grammar The correct RHS is called the handle. The most common bottom-up parsing algorithms are in the LR family (Left-to-right scan, generates Rightmost derivation) Copyright © 2006 Addison-Wesley. All rights reserved.
56
Quick Exercise Given the grammar: and the input string abbcde:
Goal -> aABe A -> Abc | b B -> d and the input string abbcde: Do a bottom-up parse to determine whether the string is a sentence in the language. Example from Engineering a Compiler Copyright © 2006 Addison-Wesley. All rights reserved.
57
Bottom-up Parsing Bottom-up parser starts with last sentential form (input sentence – no non-terminals) and produces sequence of forms until all that remains is the start symbol The parsing problem is finding the correct RHS (handle) in a right-sentential form to reduce to get the previous right-sentential form in the derivation Grammars can be left recursive Grammars do not generally include metasymbols used in EBNF Copyright © 2006 Addison-Wesley. All rights reserved.
58
Bottom-up Parsing Simple Example
Compare the derivation/parse using the same grammar Sample derivation* Sample bottom-up parse E =>E + T => E + T * F => E + T * id => E + F * id => E + id * id => T + id * id =>F + id * id => id + id * id id + id * id => F + id * id => T + id * id => E + id * id => E + F * id => E + T * id => E + T * F => E + T => E left recursion E -> E + T | T T -> T * F | F F -> (E) | id *rightmost Copyright © 2006 Addison-Wesley. All rights reserved.
59
Bottom-up Parsing - Handles
Formalized: is the handle of the right sentential form = w if and only if S =>*rm Aw =>rm w If a grammar is unambiguous, then there will be a unique handle If the grammar is ambiguous, there may be more than one possible handle =>rm is rightmost derivation leading from S (start symbol) non-terminal handle Copyright © 2006 Addison-Wesley. All rights reserved.
60
Back to our simple example
Example grammar and derivation: S -> aAc A -> aA | b S => aAc => aaAc => aabc Bottom-up parse: aabc => aaAc => aAc => S Remember there’s an incorrect choice: aabc => aaAc => aS So how can we choose the right handle? Bottom-up is based on rightmost derivation Second option is not a rightmost derivation Insight is to identify a phrase (next slide…)* Copyright © 2006 Addison-Wesley. All rights reserved. * the tool actually does this for us…
61
Phrases A phrase is a sub-sequence of a sentential form that is eventually “reduced” to a non-terminal A simple phrase can be reduced in one step The handle is the left-most simple phrase Example: S => aAc => aaAc => aabc S In this sentential form, what are the: phrases simple phrases handle A a a A c So if we have the parse tree, it’s easy to identify the handle. But how do we do it when we’re reading input from left to right? May need to delay decision, read input until find handle. sentential form b
62
Phrases – Quick Exercise
Example grammar and derivation: S -> aAC A -> aA | b C -> c S => aAC => aAc => aaAc => aabc S A In this sentential form, what are the: phrases simple phrases handle (leftmost simple phrase) a a A C sentential form b c Copyright © 2006 Addison-Wesley. All rights reserved.
63
Bottom-up Parsing (cont.)
For bottom-up parsing, need to decide: when to reduce what production (rule) to apply Shift-Reduce Algorithms Reduce is the action of replacing the handle on the top of the parse stack with its corresponding LHS Shift is the action of moving the next token to the top of the parse stack Every parser is a pushdown automaton (PDA) – can recognize a context-free grammar Copyright © 2006 Addison-Wesley. All rights reserved.
64
Bottom-up Parsing (cont.)
Knuth’s insight: A bottom-up parser could use the entire history of the parse, up to the current point, to make parsing decisions Finite, relatively small number of different possible parse situations Store history on the parse stack No need to look to left and right of substring, can just look to the left LR parsers must be constructed with a tool S A a a A c b Copyright © 2006 Addison-Wesley. All rights reserved.
65
Bottom-up Parsing (cont.)
An LR configuration stores the state of an LR parser Ss is a parser state (from a table - you’ll see) Xs is a grammar symbol (from our CFG) as are input symbols (from the string) (S0X1 S1X2 S2… XmSm, aiai+1…an$) (S0X1 S1X2 S2… A->bSm, baiai+1…an$) A reduction step is triggered when we see the symbols corresponding to a rule’s RHS on the top of the stack Copyright © 2006 Addison-Wesley. All rights reserved.
66
Structure of An LR Parser
Copyright © 2006 Addison-Wesley. All rights reserved.
67
Bottom-up Parsing (cont.)
LR parsers are table driven, where the table has two components, an ACTION table and a GOTO table The ACTION table specifies the action of the parser, given the parser state and the next token Rows are state names; columns are terminals The GOTO table specifies which state to put on top of the parse stack after a reduction action is done Rows are state names; columns are nonterminals Copyright © 2006 Addison-Wesley. All rights reserved.
68
Bottom-up Parsing (cont.)
stack input Initial configuration: (S0, a1…an$) Parser actions: (S0X1S1X2S2…SmXm, aiai+1…an$) If ACTION[Sm, ai] = Shift S, the next configuration is: (S0X1S1X2S2…SmXmSai, ai+1…an$) If ACTION[Sm, ai] = Reduce A and S = GOTO[Sm-r, A], where r = the length of , the next configuration is (S0X1S1X2S2…Xm-rSm-rAS, aiai+1…an$) new stack top new stack top was popped off stack Copyright © 2006 Addison-Wesley. All rights reserved.
69
Bottom-up Parsing (cont.)
Parser actions (continued): If ACTION[Sm, ai] = Accept, the parse is complete and no errors were found. If ACTION[Sm, ai] = Error, the parser calls an error-handling routine. Copyright © 2006 Addison-Wesley. All rights reserved.
70
LR Parsing Table E -> E + T E -> T T -> T * F T -> F
F -> (E) F -> id S4 Copyright © 2006 Addison-Wesley. All rights reserved.
71
Bottom-up Parsing/LR example
Parse id + id * id (see Chapter4Parsing.ppt, start slide 14) Stack Input Action id + id * id$ S5 0 id5 + id * id$ R6 (use GOTO[0.F]) 0 F3 R4 (use GOTO[0,T]) 0 T2 R2 (use GOTO[0,E]) 0 E1 S6 0 E1 +6 id * id$ 0 E1 +6 id5 * id$ R6 (use GOTO[6,F]) 0 E1 +6 F3 R4 (use GOTO[6,T] 0 E1 +6 T9 S7 0 E1 +6 T9 *7 id$ 0 E1 +6 T9 *7 id5 $ R6 (use GOTO[7.F]) 0 E1 +6 T9 *7 F10 R3 (use GOTO[6,T]) R1 (use GOTO[0.E]) accept Copyright © 2006 Addison-Wesley. All rights reserved.
72
Bottom-up Parsing/ LR parsers
LR parsers use a relatively small program and a parsing table Advantages of LR parsers: They will work for nearly all grammars that describe programming languages. They work on a larger class of grammars than other bottom-up algorithms, but are as efficient as any other bottom-up parser. They can detect syntax errors as soon as it is possible in left-to-right scan The LR class of grammars is a superset of the class parsable by LL parsers. Copyright © 2006 Addison-Wesley. All rights reserved.
73
Bottom-up Parsing (cont.)
A parser table can be generated from a given grammar with a tool, e.g., yacc To find out more about building the LR tables, read Engineering a Compiler or comparable compiler textbook. Copyright © 2006 Addison-Wesley. All rights reserved.
74
And now… Back to compiler tools!
Copyright © 2006 Addison-Wesley. All rights reserved.
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.