Code Optimization I: Machine Independent Optimizations Sept. 26, 2002 Topics Machine-Independent Optimizations Code motion Reduction in strength Common.

Slides:



Advertisements
Similar presentations
Code Optimization and Performance Chapter 5 CS 105 Tour of the Black Holes of Computing.
Advertisements

Program Optimization (Chapter 5)
ECE 454 Computer Systems Programming Compiler and Optimization (I) Ding Yuan ECE Dept., University of Toronto
Names and Bindings.
The Assembly Language Level
Tuan Tran. What is CISC? CISC stands for Complex Instruction Set Computer. CISC are chips that are easy to program and which make efficient use of memory.
1 Code Optimization(II). 2 Outline Understanding Modern Processor –Super-scalar –Out-of –order execution Suggested reading –5.14,5.7.
Winter 2015 COMP 2130 Intro Computer Systems Computing Science Thompson Rivers University Program Optimization.
Fundamentals of Python: From First Programs Through Data Structures
Fall 2011SYSC 5704: Elements of Computer Systems 1 SYSC 5704 Elements of Computer Systems Optimization to take advantage of hardware.
Inline Assembly Section 1: Recitation 7. In the early days of computing, most programs were written in assembly code. –Unmanageable because No type checking,
1 Lecture 6 Performance Measurement and Improvement.
Code Optimization I September 24, 2007 Topics Machine-Independent Optimizations Basic optimizations Optimization blockers class08.ppt F’07.
1 Program Optimization Professor Jennifer Rexford
Code Optimization: Machine Independent Optimizations Feb 12, 2004 Topics Machine-Independent Optimizations Machine-Dependent Opts Understanding Processor.
1 ES 314 Advanced Programming Lec 2 Sept 3 Goals: Complete the discussion of problem Review of C++ Object-oriented design Arrays and pointers.
C++ / G4MICE Course Session 3 Introduction to Classes Pointers and References Makefiles Standard Template Library.
Architecture Basics ECE 454 Computer Systems Programming
Code Optimization 1. Outline Machine-Independent Optimization –Code motion –Memory optimization Suggested reading –5.1 ~
Cache Performance Metrics
1 Code Optimization. 2 Outline Machine-Independent Optimization –Code motion –Memory optimization Suggested reading –5.2 ~ 5.6.
Recitation 7: 10/21/02 Outline Program Optimization –Machine Independent –Machine Dependent Loop Unrolling Blocking Annie Luo
University of Amsterdam Computer Systems – optimizing program performance Arnoud Visser 1 Computer Systems Optimizing program performance.
Introduction to Computer Systems Topics: Theme Five great realities of computer systems (continued) “The class that bytes”
University of Amsterdam Computer Systems – optimizing program performance Arnoud Visser 1 Computer Systems Optimizing program performance.
Code Optimization and Performance Chapters 5 and 9 perf01.ppt CS 105 “Tour of the Black Holes of Computing”
Addressing Modes Chapter 6 S. Dandamudi To be used with S. Dandamudi, “Introduction to Assembly Language Programming,” Second Edition, Springer,
Introduction to ECE 454 Computer Systems Programming Topics: Lecture topics and assignments Profiling rudiments Lab schedule and rationale Cristiana Amza.
Chapter 10 Code Optimization Zhang Jing, Wang HaiLing College of Computer Science & Technology Harbin Engineering University.
Computer Organization II Topics: Theme Five great realities of computer systems How this fits within CS curriculum CS 2733.
Compiler Optimizations ECE 454 Computer Systems Programming Topics: The Role of the Compiler Common Compiler (Automatic) Code Optimizations Cristiana Amza.
Code Optimization II: Machine Dependent Optimization Topics Machine-Dependent Optimizations Unrolling Enabling instruction level parallelism.
Machine Independent Optimizations Topics Code motion Reduction in strength Common subexpression sharing.
Code Optimization and Performance CS 105 “Tour of the Black Holes of Computing”
Programming for Performance CS 740 Oct. 4, 2000 Topics How architecture impacts your programs How (and how not) to tune your code.
Searching Topics Sequential Search Binary Search.
Embedding Assembly Code in C Programs תרגול 7 שילוב קוד אסמבלי בקוד C.
1 Code Optimization. 2 Outline Machine-Independent Optimization –Code motion –Memory optimization Suggested reading –5.2 ~ 5.6.
CS 3214 Computer Systems Godmar Back Lecture 8. Announcements Stay tuned for Project 2 & Exercise 4 Project 1 due Sep 16 Auto-fail rule 1: –Need at least.
Code Optimization and Performance I Chapter 5 perf01.ppt CS 105 “Tour of the Black Holes of Computing”
Chapter 7 Memory Management Eighth Edition William Stallings Operating Systems: Internals and Design Principles.
Measuring Time Alan L. Cox Some slides adapted from CMU slides.
©SoftMoore ConsultingSlide 1 Code Optimization. ©SoftMoore ConsultingSlide 2 Code Optimization Code generation techniques and transformations that result.
Code Optimization.
Optimization Code Optimization ©SoftMoore Consulting.
CS 3114 Many of the following slides are taken with permission from
Code Optimization I: Machine Independent Optimizations
Instructors: Dave O’Hallaron, Greg Ganger, and Greg Kesden
Program Optimization CSCE 312
Machine-Dependent Optimization
Introduction to Computer Systems
Code Optimization and Performance
Code Optimization I: Machine Independent Optimizations Feb 11, 2003
Machine-Level Programming III: Procedures Sept 18, 2001
Code Optimization(II)
Code Optimization April 6, 2000
Code Optimization I Nov. 25, 2008
Machine-Level Programming 5 Structured Data
Introduction to Computer Systems
Machine-Level Programming 5 Structured Data
COMP 2130 Intro Computer Systems Thompson Rivers University
Optimizing program performance
Cache Memories Lecture, Oct. 30, 2018
Program Optimization CSE 238/2038/2138: Systems Programming
CS 3114 Many of the following slides are taken with permission from
Machine-Independent Optimization
Lecture 11: Machine-Dependent Optimization
Code Optimization and Performance
Code Optimization II October 2, 2001
Presentation transcript:

Code Optimization I: Machine Independent Optimizations Sept. 26, 2002 Topics Machine-Independent Optimizations Code motion Reduction in strength Common subexpression sharing Tuning Identifying performance bottlenecks class10.ppt “The course that gives CMU its Zip!”

– 2 – , F’02 Great Reality #4 There’s more to performance than asymptotic complexity Constant factors matter too! Easily see 10:1 performance range depending on how code is written Must optimize at multiple levels: algorithm, data representations, procedures, and loops Must understand system to optimize performance How programs are compiled and executed How to measure program performance and identify bottlenecks How to improve performance without destroying code modularity and generality

– 3 – , F’02 Optimizing Compilers Provide efficient mapping of program to machine register allocation code selection and ordering eliminating minor inefficiencies Don’t (usually) improve asymptotic efficiency up to programmer to select best overall algorithm big-O savings are (often) more important than constant factors but constant factors also matter Have difficulty overcoming “optimization blockers” potential memory aliasing potential procedure side-effects

– 4 – , F’02 Limitations of Optimizing Compilers Operate Under Fundamental Constraint Must not cause any change in program behavior under any possible condition Often prevents it from making optimizations when would only affect behavior under pathological conditions. Behavior that may be obvious to the programmer can be obfuscated by languages and coding styles e.g., data ranges may be more limited than variable types suggest Most analysis is performed only within procedures whole-program analysis is too expensive in most cases Most analysis is based only on static information compiler has difficulty anticipating run-time inputs When in doubt, the compiler must be conservative

– 5 – , F’02 Machine-Independent Optimizations Optimizations you should do regardless of processor / compiler Code Motion Reduce frequency with which computation performed If it will always produce same result Especially moving code out of loop for (i = 0; i < n; i++) for (j = 0; j < n; j++) a[n*i + j] = b[j]; for (i = 0; i < n; i++) { int ni = n*i; for (j = 0; j < n; j++) a[ni + j] = b[j]; }

– 6 – , F’02 Compiler-Generated Code Motion Most compilers do a good job with array code + simple loop structures Code Generated by GCC for (i = 0; i < n; i++) for (j = 0; j < n; j++) a[n*i + j] = b[j]; imull %ebx,%eax# i*n movl 8(%ebp),%edi# a leal (%edi,%eax,4),%edx# p = a+i*n (scaled by 4) # Inner Loop.L40: movl 12(%ebp),%edi# b movl (%edi,%ecx,4),%eax # b+j (scaled by 4) movl %eax,(%edx)# *p = b[j] addl $4,%edx# p++ (scaled by 4) incl %ecx# j++ jl.L40 # loop if j<n for (i = 0; i < n; i++) { int ni = n*i; int *p = a+ni; for (j = 0; j < n; j++) *p++ = b[j]; }

– 7 – , F’02 Reduction in Strength Replace costly operation with simpler one Shift, add instead of multiply or divide 16*x-->x << 4 Utility machine dependent Depends on cost of multiply or divide instruction On Pentium II or III, integer multiply only requires 4 CPU cycles Recognize sequence of products for (i = 0; i < n; i++) for (j = 0; j < n; j++) a[n*i + j] = b[j]; int ni = 0; for (i = 0; i < n; i++) { for (j = 0; j < n; j++) a[ni + j] = b[j]; ni += n; }

– 8 – , F’02 Make Use of Registers Reading and writing registers much faster than reading/writing memoryLimitation Compiler not always able to determine whether variable can be held in register Possibility of Aliasing See example later

– 9 – , F’02 Machine-Independent Opts. (Cont.) Share Common Subexpressions Reuse portions of expressions Compilers often not very sophisticated in exploiting arithmetic properties /* Sum neighbors of i,j */ up = val[(i-1)*n + j]; down = val[(i+1)*n + j]; left = val[i*n + j-1]; right = val[i*n + j+1]; sum = up + down + left + right; int inj = i*n + j; up = val[inj - n]; down = val[inj + n]; left = val[inj - 1]; right = val[inj + 1]; sum = up + down + left + right; 3 multiplications: i*n, (i–1)*n, (i+1)*n1 multiplication: i*n leal -1(%edx),%ecx # i-1 imull %ebx,%ecx # (i-1)*n leal 1(%edx),%eax # i+1 imull %ebx,%eax # (i+1)*n imull %ebx,%edx # i*n

– 10 – , F’02 Vector ADT Procedures vec_ptr new_vec(int len) Create vector of specified length int get_vec_element(vec_ptr v, int index, int *dest) Retrieve vector element, store at *dest Return 0 if out of bounds, 1 if successful int *get_vec_start(vec_ptr v) Return pointer to start of vector data Similar to array implementations in Pascal, ML, Java E.g., always do bounds checking length data      012length–1

– 11 – , F’02 Optimization Example Procedure Compute sum of all elements of vector Store result at destination location void combine1(vec_ptr v, int *dest) { int i; *dest = 0; for (i = 0; i < vec_length(v); i++) { int val; get_vec_element(v, i, &val); *dest += val; }

– 12 – , F’02 Time Scales Absolute Time Typically use nanoseconds 10 –9 seconds Time scale of computer instructions Clock Cycles Most computers controlled by high frequency clock signal Typical Range 100 MHz »10 8 cycles per second »Clock period = 10ns 2 GHz »2 X 10 9 cycles per second »Clock period = 0.5ns Fish machines: 550 MHz (1.8 ns clock period)

– 13 – , F’02 Cycles Per Element Convenient way to express performance of program that operators on vectors or lists Length = n T = CPE*n + Overhead vsum1 Slope = 4.0 vsum2 Slope = 3.5

– 14 – , F’02 Optimization Example Procedure Compute sum of all elements of integer vector Store result at destination location Vector data structure and operations defined via abstract data type Pentium II/III Performance: Clock Cycles / Element (Compiled -g)31.25 (Compiled -O2) void combine1(vec_ptr v, int *dest) { int i; *dest = 0; for (i = 0; i < vec_length(v); i++) { int val; get_vec_element(v, i, &val); *dest += val; }

– 15 – , F’02 Understanding Loop Inefficiency Procedure vec_length called every iteration Even though result always the same void combine1-goto(vec_ptr v, int *dest) { int i = 0; int val; *dest = 0; if (i >= vec_length(v)) goto done; loop: get_vec_element(v, i, &val); *dest += val; i++; if (i < vec_length(v)) goto loop done: } 1 iteration

– 16 – , F’02 Move vec_length Call Out of Loop Optimization Move call to vec_length out of inner loop Value does not change from one iteration to next Code motion CPE: (Compiled -O2) vec_length requires only constant time, but significant overhead void combine2(vec_ptr v, int *dest) { int i; int length = vec_length(v); *dest = 0; for (i = 0; i < length; i++) { int val; get_vec_element(v, i, &val); *dest += val; }

– 17 – , F’02 void lower(char *s) { int i; for (i = 0; i < strlen(s); i++) if (s[i] >= 'A' && s[i] <= 'Z') s[i] -= ('A' - 'a'); } Code Motion Example #2 Procedure to Convert String to Lower Case Extracted from 213 lab submissions, Fall, 1998

– 18 – , F’02 Lower Case Conversion Performance Time quadruples when double string length Quadratic performance

– 19 – , F’02 Convert Loop To Goto Form strlen executed every iteration strlen linear in length of string Must scan string until finds '\0' Overall performance is quadratic void lower(char *s) { int i = 0; if (i >= strlen(s)) goto done; loop: if (s[i] >= 'A' && s[i] <= 'Z') s[i] -= ('A' - 'a'); i++; if (i < strlen(s)) goto loop; done: }

– 20 – , F’02 Improving Performance Move call to strlen outside of loop Since result does not change from one iteration to another Form of code motion void lower(char *s) { int i; int len = strlen(s); for (i = 0; i < len; i++) if (s[i] >= 'A' && s[i] <= 'Z') s[i] -= ('A' - 'a'); }

– 21 – , F’02 Lower Case Conversion Performance Time doubles when double string length Linear performance

– 22 – , F’02 Optimization Blocker: Procedure Calls Why couldn’t the compiler move vec_len or strlen out of the inner loop? Procedure may have side effects Alters global state each time called Function may not return same value for given arguments Depends on other parts of global state Procedure lower could interact with strlen Why doesn’t compiler look at code for vec_len or strlen ? Linker may overload with different version Unless declared static Interprocedural optimization is not used extensively due to costWarning: Compiler treats procedure call as a black box Weak optimizations in and around them

– 23 – , F’02 Reduction in Strength Optimization Avoid procedure call to retrieve each vector element Get pointer to start of array before loop Within loop just do pointer reference Not as clean in terms of data abstraction CPE: 6.00 (Compiled -O2) Procedure calls are expensive! Bounds checking is expensive void combine3(vec_ptr v, int *dest) { int i; int length = vec_length(v); int *data = get_vec_start(v); *dest = 0; for (i = 0; i < length; i++) { *dest += data[i]; }

– 24 – , F’02 Eliminate Unneeded Memory Refs Optimization Don’t need to store in destination until end Local variable sum held in register Avoids 1 memory read, 1 memory write per cycle CPE: 2.00 (Compiled -O2) Memory references are expensive! void combine4(vec_ptr v, int *dest) { int i; int length = vec_length(v); int *data = get_vec_start(v); int sum = 0; for (i = 0; i < length; i++) sum += data[i]; *dest = sum; }

– 25 – , F’02 Detecting Unneeded Memory Refs. Performance Combine3 5 instructions in 6 clock cycles addl must read and write memory Combine4 4 instructions in 2 clock cycles.L18: movl (%ecx,%edx,4),%eax addl %eax,(%edi) incl %edx cmpl %esi,%edx jl.L18 Combine3.L24: addl (%eax,%edx,4),%ecx incl %edx cmpl %esi,%edx jl.L24 Combine4

– 26 – , F’02 Optimization Blocker: Memory Aliasing Aliasing Two different memory references specify single locationExample v: [3, 2, 17] combine3(v, get_vec_start(v)+2)--> ? combine4(v, get_vec_start(v)+2)--> ?Observations Easy to have happen in C Since allowed to do address arithmetic Direct access to storage structures Get in habit of introducing local variables Accumulating within loops Your way of telling compiler not to check for aliasing

– 27 – , F’02 Machine-Independent Opt. Summary Code Motion Compilers are good at this for simple loop/array structures Don’t do well in presence of procedure calls and memory aliasing Reduction in Strength Shift, add instead of multiply or divide compilers are (generally) good at this Exact trade-offs machine-dependent Keep data in registers rather than memory compilers are not good at this, since concerned with aliasing Share Common Subexpressions compilers have limited algebraic reasoning capabilities

– 28 – , F’02 Important Tools Measurement Accurately compute time taken by code Most modern machines have built in cycle counters Using them to get reliable measurements is tricky Profile procedure calling frequencies Unix tool gprofObservation Generating assembly code Lets you see what optimizations compiler can make Understand capabilities/limitations of particular compiler

– 29 – , F’02 Code Profiling Example Task Count word frequencies in text document Produce sorted list of words from most frequent to leastSteps Convert strings to lowercase Apply hash function Read words and insert into hash table Mostly list operations Maintain counter for each unique word Sort results Data Set Collected works of Shakespeare 946,596 total words, 26,596 unique Initial implementation: 9.2 seconds 29,801the 27,529and 21,029I 20,957to 18,514of 15,370a 14010you 12,936my 11,722in 11,519that Shakespeare’s most frequent words

– 30 – , F’02 Code Profiling Augment Executable Program with Timing Functions Computes (approximate) amount of time spent in each function Time computation method Periodically (~ every 10ms) interrupt program Determine what function is currently executing Increment its timer by interval (e.g., 10ms) Also maintains counter for each function indicating number of times calledUsing gcc –O2 –pg prog. –o prog./prog Executes in normal fashion, but also generates file gmon.out gprof prog Generates profile information based on gmon.out

– 31 – , F’02 Profiling Results Call Statistics Number of calls and cumulative time for each function Performance Limiter Using inefficient sorting algorithm Single call uses 87% of CPU time % cumulative self self total time seconds seconds calls ms/call ms/call name sort_words lower find_ele_rec h_add

– 32 – , F’02 Code Optimizations First step: Use more efficient sorting function Library function qsort

– 33 – , F’02 Further Optimizations Iter first: Use iterative function to insert elements into linked list Causes code to slow down Iter last: Iterative function, places new entry at end of list Tend to place most common words at front of list Big table: Increase number of hash buckets Better hash: Use more sophisticated hash function Linear lower: Move strlen out of loop

– 34 – , F’02 Profiling Observations Benefits Helps identify performance bottlenecks Especially useful when have complex system with many componentsLimitations Only shows performance for data tested E.g., linear lower did not show big gain, since words are short Quadratic inefficiency could remain lurking in code Timing mechanism fairly crude Only works for programs that run for > 3 seconds