GC 211:Data Structures Algorithm Analysis Tools

Slides:



Advertisements
Similar presentations
Analysis of Algorithms
Advertisements

1 Data Structures Performance Analysis. 2 Fundamental Concepts Some fundamental concepts that you should know: –Dynamic memory allocation. –Recursion.
CSC401 – Analysis of Algorithms Lecture Notes 1 Introduction
© 2004 Goodrich, Tamassia 1 Lecture 01 Algorithm Analysis Topics Basic concepts Theoretical Analysis Concept of big-oh Choose lower order algorithms Relatives.
Introduction to Analysis of Algorithms
Analysis of Algorithms Algorithm Input Output. Analysis of Algorithms2 Outline and Reading Running time (§1.1) Pseudo-code (§1.1) Counting primitive operations.
Analysis of Algorithms1 CS5302 Data Structures and Algorithms Lecturer: Lusheng Wang Office: Y6416 Phone:
Analysis of Algorithms (Chapter 4)
Analysis of Algorithms1 Estimate the running time Estimate the memory space required. Time and space depend on the input size.
Analysis of Algorithms
Data Structures Performance Analysis.
Cmpt-225 Algorithm Efficiency.
Fall 2006CSC311: Data Structures1 Chapter 4 Analysis Tools Objectives –Experiment analysis of algorithms and limitations –Theoretical Analysis of algorithms.
The Seven Functions. Analysis of Algorithms. Simple Justification Techniques. 2 CPSC 3200 University of Tennessee at Chattanooga – Summer Goodrich,
Analysis of Performance
CS2210 Data Structures and Algorithms Lecture 2:
Analysis of Algorithms Algorithm Input Output © 2014 Goodrich, Tamassia, Goldwasser1Analysis of Algorithms Presentation for use with the textbook Data.
Analysis of Algorithms Lecture 2
Analysis of Algorithms
Analysis Tools Jyh-Shing Roger Jang ( 張智星 ) CSIE Dept, National Taiwan University.
Algorithm Efficiency CS 110: Data Structures and Algorithms First Semester,
Analysis of Algorithms
Analysis of Algorithms1 The Goal of the Course Design “good” data structures and algorithms Data structure is a systematic way of organizing and accessing.
Algorithm Input Output An algorithm is a step-by-step procedure for solving a problem in a finite amount of time. Chapter 4. Algorithm Analysis (complexity)
Data Structures Lecture 8 Fang Yu Department of Management Information Systems National Chengchi University Fall 2010.
1 Dr. J. Michael Moore Data Structures and Algorithms CSCE 221 Adapted from slides provided with the textbook, Nancy Amato, and Scott Schaefer.
Program Efficiency & Complexity Analysis. Algorithm Review An algorithm is a definite procedure for solving a problem in finite number of steps Algorithm.
Introduction to Analysis of Algorithms CS342 S2004.
Analysis of Algorithms Algorithm Input Output © 2010 Goodrich, Tamassia1Analysis of Algorithms.
Analysis of algorithms. What are we going to learn? Need to say that some algorithms are “better” than others Criteria for evaluation Structure of programs.
Algorithm Analysis (Big O)
Announcement We will have a 10 minutes Quiz on Feb. 4 at the end of the lecture. The quiz is about Big O notation. The weight of this quiz is 3% (please.
GC 211:Data Structures Week 2: Algorithm Analysis Tools Slides are borrowed from Mr. Mohammad Alqahtani.
1 COMP9024: Data Structures and Algorithms Week Two: Analysis of Algorithms Hui Wu Session 2, 2014
Analysis of Algorithms Algorithm Input Output © 2014 Goodrich, Tamassia, Goldwasser1Analysis of Algorithms Presentation for use with the textbook Data.
Data Structures I (CPCS-204) Week # 2: Algorithm Analysis tools Dr. Omar Batarfi Dr. Yahya Dahab Dr. Imtiaz Khan.
(Complexity) Analysis of Algorithms Algorithm Input Output 1Analysis of Algorithms.
Algorithm Analysis 1.
Analysis of Algorithms
Analysis of Algorithms
Mathematical Foundation
COMP9024: Data Structures and Algorithms
COMP9024: Data Structures and Algorithms
Analysis of Algorithms
GC 211:Data Structures Week 2: Algorithm Analysis Tools
Introduction to Algorithms
GC 211:Data Structures Algorithm Analysis Tools
Big-O notation.
03 Algorithm Analysis Hongfei Yan Mar. 9, 2016.
Analysis of Algorithms
COMP9024: Data Structures and Algorithms
Analysis of Algorithms
Analysis of Algorithms
Data Structures (CS212D) Overview & Review.
Analysis of Algorithms
Analysis of Algorithms
GC 211:Data Structures Algorithm Analysis Tools
Analysis of Algorithms
Analysis of Algorithms
Analysis of Algorithms
Analysis of Algorithms
Analysis of Algorithms
Data Structures (CS212D) Overview & Review.
Performance Evaluation
Analysis of Algorithms
CS210- Lecture 2 Jun 2, 2005 Announcements Questions
Complexity Analysis (Part II)
Analysis of Algorithms
Analysis of Algorithms
Presentation transcript:

GC 211:Data Structures Algorithm Analysis Tools Slides are borrowed from Mr. Mohammad Alqahtani and Mr. Sultan Saad Alsaleh

Algorithm Analysis: Motivation A problem can be solved in many different ways Single problem, many algorithms Which of the several algorithms should I choose? We use algorithm analysis to answer this question Writing a working program is not good enough The program may be inefficient! If the program runs on a large data set, then the running time becomes an issue

What is Algorithm Analysis? A methodology to predict the resources that the algorithm requires Computer memory Computational time We’ll focus on computational time It does not mean memory is not important Generally, there is a trade-off between the two factors Space-time trade-off is a common term

The Constant Function: For some fixed constant c, such as c = 5, c = 100 or c = 1000,000.  Regardless of the value of n, ƒ(n) will always be equal to the constant value c.  The most fundamental constant function is: g(n) = 1  Any constant function ƒ(n) = c, can be written as a constant c times g(n). That is: ƒ(n) = cg(n)

Logb a = logx a / logx b The Logarithm Function: ƒ(n) = logb n. for some constant b > 1 x = logb n if and only if bx = n  The value b is known as the base of the logarithm. The most common base for the logarithm function in computer sciences is 2. Therefore, it is typical not to write the base when it is 2: log n = log2 n  properties of logarithms: logb(xy) = logb x + logb y logb (x/y) = logb x – logb y Logb xa = alogb x Logb a = logx a / logx b

ƒ(n) = a0 + a1n + a2n2 + a3n3 + … +adnd The linear Function: ƒ(n) = n The N-Log-N Function: ƒ(n) = nlog n The Quadratic Function: ƒ(n) = n2 The Cubic Function: ƒ(n) = n3  All the function listed so far are part of a larger class of functions called the polynomial. ƒ(n) = a0 + a1n + a2n2 + a3n3 + … +adnd Where a0,a1,a2,…,ad are constants, called coefficients of the polynomial, and ad ≠ 0. the integer d, which indicates the highest power in the polynomial, is called the degree of the polynomial.

Running Time: objects into output objects. determine. best case average case worst case  Most algorithms transform input objects into output objects. 120  The running time of an algorithm typically grows with the input size. 100 Running Time 80  Average case time is often difficult to determine. 60 40  We focus on the worst case running time. Easier to analyze Crucial to applications such as games, finance and robotics 20 1000 2000 3000 4000 Input Size

Algorithm Efficiency: Growth Rate An algorithm’s time requirements can be expressed as a function of (problem) input size Problem size depends on the particular problem: For a search problem, the problem size is the number of elements in the search space For a sorting problem, the problem size is the number of elements in the given list How quickly the time of an algorithm grows as a function of problem size -- this is often called an algorithm’s growth rate

Algorithm Growth Rate Which algorithm is the most efficient? [The one with the growth rate Log N.]

Algorithmic Time Complexity Rather than counting the exact number of primitive operations, we approximate the runtime of an algorithm as a function of data size – time complexity Algorithms A, B, C and D (previous slide) belong to different complexity classes We’ll not cover complexity classes in detail We’ll briefly discuss seven basic functions which are often used in complexity analysis

Seven Basic Function Constant function f(n) = c Linear function f(n) = n Quadratic function f(n) = n2 Cubic function f(n) = n3 Log function f(n) = log n Log linear function f(n) = n log n Exponential function f(n) = bn

How to Analyse Algorithms? Experimental Approach Implement algorithms as programs and run them on computers Not a good approach, though! Results only for a limited set of test inputs Difficult comparisons due to the experiment environments (need the same computers, same operating systems, etc.) Full implementation and execution of an algorithm We need an approach which allows us to avoid experimental study

How to Analyse Algorithms? Theoretical Approach General methodology for analysing the running time Considers all possible inputs Evaluates algorithms in a way that is independent from the hardware and software environments Analyses an algorithm without implementing it Count only primitive operations used in an algorithm Associate each algorithm with a function f(n) that characterises the running time of the algorithm as a function of the input size n A good approximation of the total number of primitive operations

Primitive Operations Basic computations performed by an algorithm Each operation corresponding to a low-level instruction with a constant execution time Largely independent from the programming language Examples Evaluating an expression (x + y) Assigning a value to a variable (x ←5) Comparing two numbers (x < y) Indexing into an array (A[i]) Calling a method (mycalculator.sum()) Returning from a method (return result)

Counting Primitive Operations Total number of primitive operations executed is the running time of an algorithms is a function of the input size Example Algorithm ArrayMax(A, n) # operations currentMax ←A[0] 1 for i←1;i<n; i←i+1 do n if A[i]>currentMax then n-1 currentMax ←A[i] n-1 endif endfor return currentMax 1 Total: 3n

1 Algorithm Sum1(a[],n) - 2 { 3 S = 0.0; 4 for i=1 to n do 5 Statements S/E Freq. Total 1 Algorithm Sum1(a[],n) - 2 { 3 S = 0.0; 4 n+1 for i=1 to n do 5 n s = s+a[i]; 6 return s; 7 } 2n+3

1 Algorithm Sum2(a[],n,m) 2 { 3 for i=1 to n do; 4 for j=1 to m do 5 Statements S/E Freq. Total 1 Algorithm Sum2(a[],n,m) 2 { 3 for i=1 to n do; 4 for j=1 to m do 5 s = s+a[i][j]; 6 return s; 7 }

1 Algorithm Sum2(a[],n,m) 2 { 3 for i=1 to n do 4 5 6 return s; 7 } Statements S/E Freq. Total 1 Algorithm Sum2(a[],n,m) - 2 { 3 n+1 for i=1 to n do 4 m+1 n for j=1 to m do n(m+1) 5 m s = s+a[i][j]; nm 6 return s; 7 } 2nm+2n+2

Algorithmic Runtime Worst-case running time measures the maximum number of primitive operations executed The worst case can occur fairly often e.g. in searching a database for a particular piece of information Best-case running time measures the minimum number of primitive operations executed Finding a value in a list, where the value is at the first position Sorting a list of values, where values are already in desired order Average-case running time the efficiency averaged on all possible inputs maybe difficult to define what “average” means

Complexity Classes Suppose the execution time of algorithm A is a quadratic function of n (i.e. an2 + bn + c) Suppose the execution time of algorithm B is a linear function of n (i.e. an + b) Suppose the execution time of algorithm C is a an exponential function of n (i.e. a2n) For large problems higher order terms dominate the rest These three algorithms belong to three different “complexity classes”

Big-O and Function Growth Rate We use a convention O-notation (also called Big-Oh) to represent different complexity classes The statement “f(n) is O(g(n))” means that the growth rate of f(n) is no more than the growth rate of g(n) g(n) is an upper bound on f(n), i.e. maximum number of primitive operations We can use the big-O notation to rank functions according to their growth rate

The big-Oh notation gives an upper bound on the growth rate of a function. The statement “f(n) is O(g(n))” means that the growth rate of f(n) is no more than the growth rate of g(n). We can use the big-Oh notation to rank functions according to their growth rate.

If is f(n) a polynomial of degree d, then f(n) is O(nd), i.e., Drop lower-order terms 2.Drop constant factors Use the smallest possible class of functions Say “2n is O(n)” instead of “2n is O(n2)”. Use the simplest expression of the class Say “3n + 5 is O(n)” instead of “3n + 5 is O(3n)”

Big-O: Functions Ranking BETTER WORSE O(1) constant time O(log n) log time O(n) linear time O(n log n) log linear time O(n2) quadratic time O(n3) cubic time O(2n) exponential time

O(1) Constant O(log n) Logarithmic O(n) Linear O(n log n) n log n O(nc) O(n2), O(n3), O(n16), etc Polynomial O(cn) O(1.6n), O(2n), O(3n), etc Exponential O(n!) Factorial O(nn) n power n

Heap sort, Quick sort (average), Merge sort Push, Pop, Enqueue (if there is a tail reference), Dequeue, Accessing an array element O(log n) Binary search O(n) Linear search O(n log n) Heap sort, Quick sort (average), Merge sort O(n2) Selection sort, Insertion sort, Bubble sort O(n3) Matrix multiplication O(2n) Towers of Hanoi O(n!) All permutation of N elements O(nn) ….

Simplifications Keep just one term No constant coefficients are kept the fastest growing term (dominates the runtime) No constant coefficients are kept Constant coefficients affected by machines, languages, etc Asymptotic behavior (as n gets large) is determined entirely by the dominating term Example: T(n) = 10 n3 + n2 + 40n + 800 If n = 1,000, then T(n) = 10,001,040,800 error is 0.01% if we drop all but the n3 (the dominating) term

Big Oh: Some Examples n3 – 3n = O(n3) 1 + 4n = O(n)

7n-2 is O(n) 2. 3n3 + 20n2 + 5 is O(n3) 3. 3 log n + 5 is O(log n) need c > 0 and n0  1 such that 7n-2  cn for n  n0 this is true for c = 7 and n0 = 1 3n3 + 20n2 + 5 2. 3n3 + 20n2 + 5 is O(n3) need c > 0 and n0  1 such that 3n3 + 20n2 + 5  cn3 for n  n0 this is true for c = 4 and n0 = 21 3 log n + 5 3. 3 log n + 5 is O(log n) need c > 0 and n0  1 such that 3 log n + 5  clog n for n  n0 this is true for c = 8 and n0 = 2

Example

Asymptotic Algorithm Analysis Determines the running time in big-O notation Asymptotic analysis find the worst-case number of primitives operations executed as a function of the input size express this function with big-O notation Example: algorithm arrayMax executes at most 7n − 2primitive operations algorithm arrayMax runs in O(n) time

Practice: Express the following functions in terms of Big-O notation (a, b and c are constants) f(n) = an2 + bn + c f(n) = 2n + n log n + c f(n) = n log n + b log n + c

a). s = 0; for (i = 1; i < n-1; i++) 1 n-1 n-2 Total: 2n-2 O(n) d). i = 0; while (i <= 10) i = i + 1; 1 12 11 Total: 24 O(1)

You can find more here : https://www.youtube.com/watch?v=sblr6SXgyLA