Code Shape II Expressions & Arrays Copyright 2003, Keith D. Cooper, Ken Kennedy & Linda Torczon, all rights reserved.

Slides:



Advertisements
Similar presentations
Instruction Set Design
Advertisements

Optimizing Compilers for Modern Architectures Allen and Kennedy, Chapter 13 Compiling Array Assignments.
Intermediate Code Generation
Programming Languages and Paradigms
1 Compiler Construction Intermediate Code Generation.
Chapter 5: Elementary Data Types Properties of types and objects –Data objects, variables and constants –Data types –Declarations –Type checking –Assignment.
Chapter Four Data Types Pratt 2 Data Objects A run-time grouping of one or more pieces of data in a virtual machine a container for data it can be –system.
Chapter 6 Structured Data Types Arrays Records. Copyright © 2007 Addison-Wesley. All rights reserved. 1–2 Definitions data type –collection of data objects.
Introduction to Code Optimization Comp 412 Copyright 2010, Keith D. Cooper & Linda Torczon, all rights reserved. Students enrolled in Comp 412 at Rice.
Introduction to Code Generation Comp 412 Copyright 2010, Keith D. Cooper & Linda Torczon, all rights reserved. Students enrolled in Comp 412 at Rice University.
Code Shape, Part II Addressing Arrays, Aggregates, & Strings Comp 412 Copyright 2010, Keith D. Cooper & Linda Torczon, all rights reserved. Students enrolled.
Intermediate code generation. Code Generation Create linear representation of program Result can be machine code, assembly code, code for an abstract.
Code Shape IV Procedure Calls & Dispatch Copyright 2003, Keith D. Cooper, Ken Kennedy & Linda Torczon, all rights reserved. Students enrolled in Comp 412.
Intermediate Representations Copyright 2003, Keith D. Cooper, Ken Kennedy & Linda Torczon, all rights reserved. Students enrolled in Comp 412 at Rice University.
Code Shape III Booleans, Relationals, & Control flow Copyright 2003, Keith D. Cooper, Ken Kennedy & Linda Torczon, all rights reserved. Students enrolled.
1 Chapter 7: Runtime Environments. int * larger (int a, int b) { if (a > b) return &a; //wrong else return &b; //wrong } int * larger (int *a, int *b)
1 Handling nested procedures Method 1 : static (access) links –Reference to the frame of the lexically enclosing procedure –Static chains of such links.
The Procedure Abstraction Part III: Allocating Storage & Establishing Addressability Copyright 2003, Keith D. Cooper, Ken Kennedy & Linda Torczon, all.
4/23/09Prof. Hilfinger CS 164 Lecture 261 IL for Arrays & Local Optimizations Lecture 26 (Adapted from notes by R. Bodik and G. Necula)
ISBN Chapter 6 Data Types: Structured types.
Cse321, Programming Languages and Compilers 1 6/19/2015 Lecture #18, March 14, 2007 Syntax directed translations, Meanings of programs, Rules for writing.
Run-Time Storage Organization
Instruction Selection, II Tree-pattern matching Copyright 2003, Keith D. Cooper, Ken Kennedy & Linda Torczon, all rights reserved. Students enrolled in.
Context-sensitive Analysis, II Ad-hoc syntax-directed translation, Symbol Tables, andTypes.
Introduction to Optimization Copyright 2003, Keith D. Cooper, Ken Kennedy & Linda Torczon, all rights reserved.
Arrays Data Structures - structured data are data organized to show the relationship among the individual elements. It usually requires a collecting mechanism.
Chapter 6 Structured Data Types Arrays Records. Copyright © 2007 Addison-Wesley. All rights reserved. 1–2 Definitions data type –collection of data objects.
5.3 Machine-Independent Compiler Features
Chapter 7: Runtime Environment –Run time memory organization. We need to use memory to store: –code –static data (global variables) –dynamic data objects.
Imperative Programming
Operator Precedence First the contents of all parentheses are evaluated beginning with the innermost set of parenthesis. Second all multiplications, divisions,
IT253: Computer Organization Lecture 4: Instruction Set Architecture Tonga Institute of Higher Education.
CSC3315 (Spring 2009)1 CSC 3315 Programming Languages Hamid Harroud School of Science and Engineering, Akhawayn University
Names Variables Type Checking Strong Typing Type Compatibility 1.
Runtime Environments Compiler Construction Chapter 7.
Chapter 12 Recursion, Complexity, and Searching and Sorting
Compiler Construction
The Procedure Abstraction, Part VI: Inheritance in OOLs Comp 412 Copyright 2010, Keith D. Cooper & Linda Torczon, all rights reserved. Students enrolled.
The Procedure Abstraction, Part V: Support for OOLs Comp 412 Copyright 2010, Keith D. Cooper & Linda Torczon, all rights reserved. Students enrolled in.
Compiler Chapter# 5 Intermediate code generation.
Copyright © 2005 Elsevier Chapter 8 :: Subroutines and Control Abstraction Programming Language Pragmatics Michael L. Scott.
Basic Semantics Associating meaning with language entities.
1 Records Record aggregate of data elements –Possibly heterogeneous –Elements/slots are identified by names –Elements in same fixed order in all records.
Programming Languages and Paradigms Imperative Programming.
Code Shape IV Procedure Calls & Dispatch Copyright 2003, Keith D. Cooper, Ken Kennedy & Linda Torczon, all rights reserved. Students enrolled in Comp 412.
Addressing Modes Chapter 6 S. Dandamudi To be used with S. Dandamudi, “Introduction to Assembly Language Programming,” Second Edition, Springer,
Introduction to Code Generation Copyright 2003, Keith D. Cooper, Ken Kennedy & Linda Torczon, all rights reserved. Students enrolled in Comp 412 at Rice.
Chapter 7 Object Code Generation. Chapter 7 -- Object Code Generation2  Statements in 3AC are simple enough that it is usually no great problem to map.
Code Shape II Expressions & Assignment Copyright 2003, Keith D. Cooper, Ken Kennedy & Linda Torczon, all rights reserved. Students enrolled in Comp 412.
Procedures and Functions Procedures and Functions – subprograms – are named fragments of program they can be called from numerous places  within a main.
1 ENERGY 211 / CME 211 Lecture 4 September 29, 2008.
Intermediate Code Generation CS 671 February 14, 2008.
7-Nov Fall 2001: copyright ©T. Pearce, D. Hutchinson, L. Marshall Oct lecture23-24-hll-interrupts 1 High Level Language vs. Assembly.
Run-Time Environments Presented By: Seema Gupta 09MCA102.
Introduction to Optimization
The Procedure Abstraction Part IV: Allocating Storage & Establishing Addressability Copyright 2003, Keith D. Cooper, Ken Kennedy & Linda Torczon, all rights.
Chapter 9 :: Subroutines and Control Abstraction
Introduction to Optimization
Intermediate Representations
Introduction to Code Generation
Chapter 6 Intermediate-Code Generation
Code Shape III Booleans, Relationals, & Control flow
Instruction Selection, II Tree-pattern matching
Intermediate Representations
Code Shape IV Procedure Calls & Dispatch
Code Shape II Expressions & Assignment
Lecture 15: Code Generation - Instruction Selection
Introduction to Optimization
8 Code Generation Topics A simple code generator algorithm
Procedure Linkages Standard procedure linkage Procedure has
Presentation transcript:

Code Shape II Expressions & Arrays Copyright 2003, Keith D. Cooper, Ken Kennedy & Linda Torczon, all rights reserved.

Code Shape What if x is 2 and z is 3? What if y+z is evaluated earlier? The “best” shape for x+y+z depends on contextual knowledge  There may be several conflicting options x + y + z x + y  t1 t1+ z  t2 x + z  t1 t1+ y  t2 y + z  t1 t1+ z  t2  zyx   z yx   y zx   x zy Addition is commutative & associative for integers

Code Shape Another example -- the case statement Implement it as cascaded if-then-else statements  Cost depends on where your case actually occurs  O(number of cases) Implement it as a binary search  Need a dense set of conditions to search  Uniform (log n) cost Implement it as a jump table  Lookup address in a table & jump to it  Uniform (constant) cost Compiler must choose best implementation strategy No amount of massaging or transforming will convert one into another

Generating Code for Expressions The key code quality issue is holding values in registers When can a value be safely allocated to a register?  When only 1 name can reference its value  Pointers, parameters, aggregates & arrays all cause trouble When should a value be allocated to a register?  When it is both safe & profitable Encoding this knowledge into the IR Use code shape to make it known to every later phase Assign a virtual register to anything that can go into one Load or store the others at each reference Relies on a strong register allocator

Generating Code for Expressions The concept Use a simple treewalk evaluator Bury complexity in routines it calls > base(), offset(), & val() Implements expected behavior > Visits & evaluates children > Emits code for the op itself > Returns register with result Works for simple expressions Easily extended to other operators Does not handle control flow expr(node) { int result, t1, t2; switch (type(node)) { case , , ,  : t1  expr(left child(node)); t2  expr(right child(node)); result  NextRegister(); emit (op(node), t1, t2, result); break; case IDENTIFIER: t1  base(node); t2  offset(node); result  NextRegister(); emit (loadAO, t1, t2, result); break; case NUMBER: result  NextRegister(); emit (loadI, val(node), none, result); break; } return result; }

Generating Code for Expressions Example: Produces:  xy expr(node) { int result, t1, t2; switch (type(node)) { case , , ,  : t1  expr(left child(node)); t2  expr(right child(node)); result  NextRegister(); emit (op(node), t1, t2, result); break; case IDENTIFIER: t1  base(node); t2  offset(node); result  NextRegister(); emit (loadAO, t1, t2, result); break; case NUMBER: result  NextRegister(); emit (loadI, val(node), none, result); break; } return result; }

Generating Code for Expressions Example: Generates:   x y 2 expr(node) { int result, t1, t2; switch (type(node)) { case , , ,  : t1  expr(left child(node)); t2  expr(right child(node)); result  NextRegister(); emit (op(node), t1, t2, result); break; case IDENTIFIER: t1  base(node); t2  offset(node); result  NextRegister(); emit (loadAO, t1, t2, result); break; case NUMBER: result  NextRegister(); emit (loadI, val(node), none, result); break; } return result; }

Extending the Simple Treewalk Algorithm More complex cases for IDENTIFIER What about values in registers?  Modify the IDENTIFIER case  Already in a register  return the register name  Not in a register  load it as before, but record the fact  Choose names to avoid creating false dependences What about parameter values?  Many linkages pass the first several values in registers  Call-by-value  just a local variable with “funny” offset  Call-by-reference  needs an extra indirection What about function calls in expressions?  Generate the calling sequence & load the return value  Severely limits compiler’s ability to reorder operations

Extending the Simple Treewalk Algorithm Adding other operators Evaluate the operands, then perform the operation Complex operations may turn into library calls Handle assignment as an operator Mixed-type expressions Insert conversions as needed from conversion table Most languages have symmetric & rational conversion tables Typical Addition Table

Handling Assignment (just another operator) lhs  rhs Strategy Evaluate rhs to a value (an rvalue) Evaluate lhs to a location (an lvalue)  lvalue is a register  move rhs  lvalue is an address  store rhs If rvalue & lvalue have different types  Evaluate rvalue to its “natural” type  Convert that value to the type of *lvalue Unambiguous scalars go into registers Ambiguous scalars or aggregates go into memory Let hardware sort out the addresses !

Handling Assignment What if the compiler cannot determine the rhs’s type ? This is a property of the language & the specific program If type-safety is desired, compiler must insert a run-time check Add a tag field to the data items to hold type information Code for assignment becomes more complex evaluate rhs if type(lhs)  rhs.tag then convert rhs to type(lhs) or signal a run-time error lhs  rhs This is much more complex than if it knew the types

Handling Assignment Compile-time type-checking Goal is to eliminate both the check & the tag Determine, at compile time, the type of each subexpression Use compile-time types to determine if a run-time check is needed Optimization strategy If compiler knows the type, move the check to compile-time Unless tags are needed for garbage collection, eliminate them If check is needed, try to overlap it with other computation Can design the language so all checks are static

How does the compiler handle A [i,j] ? First, must agree on a storage scheme Row-major order (most languages) Lay out as a sequence of consecutive rows Rightmost subscript varies fastest A[1,1], A[1,2], A[1,3], A[2,1], A[2,2], A[2,3] Column-major order (Fortran) Lay out as a sequence of columns Leftmost subscript varies fastest A[1,1], A[2,1], A[1,2], A[2,2], A[1,3], A[2,3] Indirection vectors (Java) Vector of pointers to pointers to … to values Takes much more space, trades indirection for arithmetic Not amenable to analysis

Laying Out Arrays The Concept Row-major order Column-major order Indirection vectors 1,11,21,31,42,12,22,32,4 A 1,12,11,22,21,32,31,42,4 A 1,11,21,31,4 2,12,22,32,4 A 1,11,21,31,4 2,12,22,32,4 A These have distinct & different cache behavior

Computing an Array Address A[ i + ( i – low ) x sizeof(A[1]) In general: base(A) + ( i – low ) x sizeof(A[1]) Almost always a power of 2, known at compile-time  use a shift for speed

Computing an Array Address A[ i + ( i – low ) x sizeof(A[1]) In general: base(A) + ( i – low ) x sizeof(A[1]) Almost always a power of 2, known at compile-time  use a shift for speed int A[1:10]  low is 1 Make low 0 for faster access (saves a – )

Computing an Array Address A[ i + ( i – low ) x sizeof(A[1]) In general: base(A) + ( i – low ) x sizeof(A[1]) What about A[i 1,i 2 ] ? Row-major order, two + (( i 1 – low 1 ) x (high 2 – low 2 + 1) + i 2 – low 2 ) x sizeof(A[1]) Column-major order, two + (( i 2 – low 2 ) x (high 1 – low 1 + 1) + i 1 – low 1 ) x sizeof(A[1]) Indirection vectors, two dimensions *(A[i 1 ])[i 2 ] — where A[i 1 ] is, itself, a 1-d array reference This stuff looks expensive! Lots of implicit +, -, x ops

where w = sizeof(A[1,1]) Optimizing Address Calculation for A [i,j] In row-major + (i–low 1 )(high 2 –low 2 +1) x w + (j – low 2 ) x w Which can be factored + i x (high 2 –low 2 +1) x w + j x w – (low 1 x (high 2 –low 2 +1) x w) + (low 2 x w) If low i, high i, and w are known, the last term is a constant 0 – (low 1 x (high 2 –low 2 +1) x w) + (low 2 x w) And len 2 as (high 2 -low 2 +1) Then, the address expression 0 + (i x len 2 + j ) x w Compile-time constants

Array References What about arrays as actual parameters? Whole arrays, as call-by-reference parameters Need dimension information  build a dope vector Store the values in the calling sequence Pass the address of the dope vector in the parameter slot Generate complete address polynomial at each reference What about call-by-value? Most c-b-v languages pass arrays by reference This is a language design low 1 high 1 low 2 high 2

Array References What about A[12] as an actual parameter? If corresponding parameter is a scalar, it’s easy Pass the address or value, as needed Must know about both formal & actual parameter Language definition must force this interpretation What if corresponding parameter is an array? Must know about both formal & actual parameter Meaning must be well-defined and understood Cross-procedural checking of conformability  Again, we’re treading on language design issues

Example: Array Address Calculations in a Loop DO J = 1, N A[I,J] = A[I,J] + B[I,J] END DO Naïve: Perform the address calculation twice DO J = 1, N R1 0 + (J x len 1 + I ) x floatsize R2 0 + (J x len 1 + I ) x floatsize MEM(R1) = MEM(R1) + MEM(R2) END DO

Example: Array Address Calculations in a Loop DO J = 1, N A[I,J] = A[I,J] + B[I,J] END DO Sophisticated: Move comon calculations out of loop R1 = I x floatsize c = len 1 x floatsize ! Compile-time constant R2 0 + R1 R3 0 + R1 DO J = 1, N a = J x c R4 = R2 + a R5 = R3 + a MEM(R4) = MEM(R4) + MEM(R5) END DO