Recap from last time We were trying to do Common Subexpression Elimination Compute expressions that are available at each program point.

Slides:



Advertisements
Similar presentations
SSA and CPS CS153: Compilers Greg Morrisett. Monadic Form vs CFGs Consider CFG available exp. analysis: statement gen's kill's x:=v 1 p v 2 x:=v 1 p v.
Advertisements

Data-Flow Analysis II CS 671 March 13, CS 671 – Spring Data-Flow Analysis Gather conservative, approximate information about what a program.
School of EECS, Peking University “Advanced Compiler Techniques” (Fall 2011) SSA Guo, Yao.
P3 / 2004 Register Allocation. Kostis Sagonas 2 Spring 2004 Outline What is register allocation Webs Interference Graphs Graph coloring Spilling Live-Range.
Chapter 9 Code optimization Section 0 overview 1.Position of code optimizer 2.Purpose of code optimizer to get better efficiency –Run faster –Take less.
Context-Sensitive Interprocedural Points-to Analysis in the Presence of Function Pointers Presentation by Patrick Kaleem Justin.
Course Outline Traditional Static Program Analysis –Theory Compiler Optimizations; Control Flow Graphs Data-flow Analysis – today’s class –Classic analyses.
Control-Flow Graphs & Dataflow Analysis CS153: Compilers Greg Morrisett.
CS412/413 Introduction to Compilers Radu Rugina Lecture 37: DU Chains and SSA Form 29 Apr 02.
Register Allocation CS 671 March 27, CS 671 – Spring Register Allocation - Motivation Consider adding two numbers together: Advantages: Fewer.
Components of representation Control dependencies: sequencing of operations –evaluation of if & then –side-effects of statements occur in right order Data.
Program Representations. Representing programs Goals.
Example in SSA X := Y op Z in out F X := Y op Z (in) = in [ { X ! Y op Z } X :=  (Y,Z) in 0 out F X :=   (in 0, in 1 ) = (in 0 Å in 1 ) [ { X ! E |
Common Sub-expression Elim Want to compute when an expression is available in a var Domain:
Course project presentations No midterm project presentation Instead of classes, next week I’ll meet with each group individually, 30 mins each Two time.
Representing programs Goals. Representing programs Primary goals –analysis is easy and effective just a few cases to handle directly link related things.
Feedback: Keep, Quit, Start
Recap from last time Saw several examples of optimizations –Constant folding –Constant Prop –Copy Prop –Common Sub-expression Elim –Partial Redundancy.
Next Section: Pointer Analysis Outline: –What is pointer analysis –Intraprocedural pointer analysis –Interprocedural pointer analysis (Wilson & Lam) –Unification.
Interprocedural analyses and optimizations. Costs of procedure calls Up until now, we treated calls conservatively: –make the flow function for call nodes.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Emery Berger University of Massachusetts, Amherst Advanced Compilers CMPSCI 710.
Administrative info Subscribe to the class mailing list –instructions are on the class web page, click on “Course Syllabus” –if you don’t subscribe by.
1 Intermediate representation Goals: –encode knowledge about the program –facilitate analysis –facilitate retargeting –facilitate optimization scanning.
Global optimization. Data flow analysis To generate better code, need to examine definitions and uses of variables beyond basic blocks. With use- definition.
Administrative info Subscribe to the class mailing list –instructions are on the class web page, which is accessible from my home page, which is accessible.
Inlining pros and cons (discussion). Inlining pros and cons Pros –eliminate overhead of call/return sequence –eliminate overhead of passing args & returning.
Previous finals up on the web page use them as practice problems look at them early.
Another example p := &x; *p := 5 y := x + 1;. Another example p := &x; *p := 5 y := x + 1; x := 5; *p := 3 y := x + 1; ???
From last class: CSE Want to compute when an expression is available in a var Domain:
Loop invariant detection using SSA An expression is invariant in a loop L iff: (base cases) –it’s a constant –it’s a variable use, all of whose single.
CS 412/413 Spring 2007Introduction to Compilers1 Lecture 29: Control Flow Analysis 9 Apr 07 CS412/413 Introduction to Compilers Tim Teitelbaum.
Class canceled next Tuesday. Recap: Components of IR Control dependencies: sequencing of operations –evaluation of if & then –side-effects of statements.
1 Copy Propagation What does it mean? – Given an assignment x = y, replace later uses of x with uses of y, provided there are no intervening assignments.
Common Sub-expression Elim Want to compute when an expression is available in a var Domain:
From last time: Inlining pros and cons Pros –eliminate overhead of call/return sequence –eliminate overhead of passing args & returning results –can optimize.
Projects. Dataflow analysis Dataflow analysis: what is it? A common framework for expressing algorithms that compute information about a program Why.
Recap from last time: live variables x := 5 y := x + 2 x := x + 1 y := x y...
Course project presentations Midterm project presentation Originally scheduled for Tuesday Nov 4 th Can move to Th Nov 6 th or Th Nov 13 th.
Machine-Independent Optimizations Ⅰ CS308 Compiler Theory1.
Global optimization. Data flow analysis To generate better code, need to examine definitions and uses of variables beyond basic blocks. With use- definition.
Schedule Midterm out tomorrow, due by next Monday Final during finals week Project updates next week.
Direction of analysis Although constraints are not directional, flow functions are All flow functions we have seen so far are in the forward direction.
PSUCS322 HM 1 Languages and Compiler Design II IR Code Optimization Material provided by Prof. Jingke Li Stolen with pride and modified by Herb Mayer PSU.
Recap from last time We saw various different issues related to program analysis and program transformations You were not expected to know all of these.
From last class. The above is Click’s solution (PLDI 95)
Precision Going back to constant prop, in what cases would we lose precision?
Compiler Principles Fall Compiler Principles Lecture 0: Local Optimizations Roman Manevich Ben-Gurion University.
Program Representations. Representing programs Goals.
Interprocedural analyses and optimizations. Costs of procedure calls Up until now, we treated calls conservatively: –make the flow function for call nodes.
CS412/413 Introduction to Compilers and Translators April 2, 1999 Lecture 24: Introduction to Optimization.
Interprocedural analyses and optimizations. Costs of procedure calls Up until now, we treated calls conservatively: –make the flow function for call nodes.
LECTURE 19 Subroutines and Parameter Passing. ABSTRACTION Recall: Abstraction is the process by which we can hide larger or more complex code fragments.
CS 412/413 Spring 2005Introduction to Compilers1 CS412/CS413 Introduction to Compilers Tim Teitelbaum Lecture 30: Loop Optimizations and Pointer Analysis.
Code Optimization Overview and Examples
Dataflow analysis.
Program Representations
Fall Compiler Principles Lecture 8: Loop Optimizations
Topic 10: Dataflow Analysis
Program Slicing Baishakhi Ray University of Virginia
Code Optimization Overview and Examples Control Flow Graph
Static Single Assignment Form (SSA)
Optimizations using SSA
Fall Compiler Principles Lecture 10: Loop Optimizations
Pointer analysis.
EECS 583 – Class 7 Static Single Assignment Form
Interprocedural analyses and optimizations
EECS 583 – Class 7 Static Single Assignment Form
SSA-based Optimizations
CSE P 501 – Compilers SSA Hal Perkins Autumn /31/2019
Presentation transcript:

Recap from last time We were trying to do Common Subexpression Elimination Compute expressions that are available at each program point

Recap from last time X := Y op Z in out F X := Y op Z (in) = in – { X ! * } – { * !... X... } [ { X ! Y op Z | X  Y Æ X  Z} X := Y in out F X := Y (in) = in – { X ! * } – { * !... X... } [ { X ! E | Y ! E 2 in }

Example

Problems z := j * 4 is not optimized to z := x, even though x contains the value j * 4 m := b + a is not optimized to m := i, even i contains a + b w := 4 * m it not optimized to w := x, even though x contains the value 4 *m

Problems: more abstractly Available expressions overly sensitive to name choices, operand orderings, renamings, assignments Use SSA: distinct values have distinct names Do copy prop before running available exprs Adopt canonical form for commutative ops

Example in SSA X := Y op Z in out F X := Y op Z (in) = X :=  (Y,Z) in 0 out F X := Y (in 0, in 1 ) = in 1

Example in SSA X := Y op Z in out F X := Y op Z (in) = in [ { X ! Y op Z } X :=  (Y,Z) in 0 out F X := Y (in 0, in 1 ) = (in 0 Å in 1 ) [ { X ! E | Y ! E 2 in 0 Æ Z ! E 2 in 1 } in 1

Example in SSA

What about pointers?

Option 1: don’t use SSA for point-to memory Option 2: insert copies between SSA vars and real vars

SSA helps us with CSE Let’s see what else SSA can help us with Loop-invariant code motion

Two steps: analysis and transformations Step1: find invariant computations in loop –invariant: computes same result each time evaluated Step 2: move them outside loop –to top if used within loop: code hoisting –to bottom if used after loop: code sinking

Example

Detecting loop invariants An expression is invariant in a loop L iff: (base cases) –it’s a constant –it’s a variable use, all of whose defs are outside of L (inductive cases) –it’s a pure computation all of whose args are loop- invariant –it’s a variable use with only one reaching def, and the rhs of that def is loop-invariant

Computing loop invariants Option 1: iterative dataflow analysis –optimistically assume all expressions loop-invariant, and propagate Option 2: build def/use chains –follow chains to identify and propagate invariant expressions Option 3: SSA –like option 2, but using SSA instead of ef/use chains

Example using def/use chains An expression is invariant in a loop L iff: (base cases) –it’s a constant –it’s a variable use, all of whose defs are outside of L (inductive cases) –it’s a pure computation all of whose args are loop-invariant –it’s a variable use with only one reaching def, and the rhs of that def is loop-invariant

Example using def/use chains An expression is invariant in a loop L iff: (base cases) –it’s a constant –it’s a variable use, all of whose defs are outside of L (inductive cases) –it’s a pure computation all of whose args are loop-invariant –it’s a variable use with only one reaching def, and the rhs of that def is loop-invariant

Loop invariant detection using SSA An expression is invariant in a loop L iff: (base cases) –it’s a constant –it’s a variable use, all of whose single defs are outside of L (inductive cases) –it’s a pure computation all of whose args are loop- invariant –it’s a variable use whose single reaching def, and the rhs of that def is loop-invariant  functions are not pure

Example using SSA An expression is invariant in a loop L iff: (base cases) –it’s a constant –it’s a variable use, all of whose single defs are outside of L (inductive cases) –it’s a pure computation all of whose args are loop-invariant –it’s a variable use whose single reaching def, and the rhs of that def is loop-invariant  functions are not pure

Example using SSA and preheader An expression is invariant in a loop L iff: (base cases) –it’s a constant –it’s a variable use, all of whose single defs are outside of L (inductive cases) –it’s a pure computation all of whose args are loop-invariant –it’s a variable use whose single reaching def, and the rhs of that def is loop-invariant  functions are not pure

Code motion

Example

Lesson from example: domination restriction

Domination restriction in for loops

Avoiding domination restriction

Another example

Data dependence restriction

Avoiding data restriction

Representing control dependencies We’ve seen SSA, a way to encode data dependencies better than just def/use chains –makes CSE easier –makes loop invariant detection easier –makes code motion easier Now we move on to looking at how to encode control dependencies

Control Dependences A node (basic block) Y is control-dependent on another X iff X determines whether Y executes –there exists a path from X to Y s.t. every node in the path other than X and Y is post-dominated by Y –X is not post-dominated by Y

Control Dependences A node (basic block) Y is control-dependent on another X iff X determines whether Y executes –there exists a path from X to Y s.t. every node in the path other than X and Y is post-dominated by Y –X is not post-dominated by Y

Example

Control Dependence Graph Control dependence graph: Y descendent of X iff Y is control dependent on X –label each child edge with required condition –group all children with same condition under region node Program dependence graph: super-impose dataflow graph (in SSA form or not) on top of the control dependence graph

Example

Another example

Summary so far Dataflow analysis –flow functions, lattice theoretic framework, optimistic iterative analysis, MOP Advanced Program Representations –SSA, CDG, PDG, code motion Next: dealing with procedures

Interprocedural analyses and optimizations

Costs of procedure calls Up until now, we treated calls conservatively: –make the flow function for call nodes return top –start iterative analysis with incoming edge of the CFG set to top –This leads to less precise results: “lost-precision” cost Calls also incur a direct runtime cost –cost of call, return, argument & result passing, stack frame maintainance –“direct runtime” cost

Addressing costs of procedure calls Technique 1: try to get rid of calls, using inlining and other techniques Technique 2: interprocedural analysis, for calls that are left

Inlining Replace call with body of callee Turn parameter- and result-passing into assignments –do copy prop to eliminate copies Manage variable scoping correctly –rename variables where appropriate

Program representation for inlining Call graph –nodes are procedures –edges are calls, labelled by invocation counts/frequency Hard cases for builing call graph –calls to/from external routines –calls through pointers, function values, messages Where in the compiler should inlining be performed?

Inlining pros and cons (discussion)

Inlining pros and cons Pros –eliminate overhead of call/return sequence –eliminate overhead of passing args & returning results –can optimize callee in context of caller and vice versa Cons –can increase compiled code space requirements –can slow down compilation –recursion? Virtual inlining: simulate inlining during analysis of caller, but don’t actually perform the inlining

Which calls to inline (discussion) What affects the decision as to which calls to inline?

Which calls to inline What affects the decision as to which calls to inline? –size of caller and callee (easy to compute size before inlining, but what about size after inlining?) –frequency of call (static estimates or dynamic profiles) –call sites where callee benefits most from optimization (not clear how to quantify) –programmer annotations (if so, annotate procedure or call site? Also, should the compiler really listen to the programmer?)