Design and Analysis of Algorithms. Part 1 Program Costs and Asymptotic Notations

Similar documents
Growth of Functions (CLRS 2.3,3)

Analysis of Algorithm Efficiency. Dr. Yingwu Zhu

CSC Design and Analysis of Algorithms. Lecture 1

Module 1: Analyzing the Efficiency of Algorithms

Ch01. Analysis of Algorithms

Big O (Asymptotic Upper Bound)

Ch 01. Analysis of Algorithms

Big O 2/14/13. Administrative. Does it terminate? David Kauchak cs302 Spring 2013

Lecture 2. Fundamentals of the Analysis of Algorithm Efficiency

Lecture 2: Asymptotic Notation CSCI Algorithms I

Lecture 2. More Algorithm Analysis, Math and MCSS By: Sarah Buchanan

Module 1: Analyzing the Efficiency of Algorithms

Algorithms Design & Analysis. Analysis of Algorithm

Data Structures and Algorithms Chapter 2

Algorithms and Their Complexity

csci 210: Data Structures Program Analysis

Input Decidable Language -- Program Halts on all Input Encoding of Input -- Natural Numbers Encoded in Binary or Decimal, Not Unary

Computational Complexity

5 + 9(10) + 3(100) + 0(1000) + 2(10000) =

csci 210: Data Structures Program Analysis

Principles of Algorithm Analysis

Asymptotic Notation. such that t(n) cf(n) for all n n 0. for some positive real constant c and integer threshold n 0

Analysis of Algorithms [Reading: CLRS 2.2, 3] Laura Toma, csci2200, Bowdoin College

COMP 382: Reasoning about algorithms

Data Structures and Algorithms Running time and growth functions January 18, 2018

Lecture 5: Loop Invariants and Insertion-sort

Algorithm efficiency can be measured in terms of: Time Space Other resources such as processors, network packets, etc.

Running Time Evaluation

Lecture 1 - Preliminaries

Reminder of Asymptotic Notation. Inf 2B: Asymptotic notation and Algorithms. Asymptotic notation for Running-time

Define Efficiency. 2: Analysis. Efficiency. Measuring efficiency. CSE 417: Algorithms and Computational Complexity. Winter 2007 Larry Ruzzo

CS 4407 Algorithms Lecture 2: Iterative and Divide and Conquer Algorithms

Defining Efficiency. 2: Analysis. Efficiency. Measuring efficiency. CSE 421: Intro Algorithms. Summer 2007 Larry Ruzzo

Algorithm. Executing the Max algorithm. Algorithm and Growth of Functions Benchaporn Jantarakongkul. (algorithm) ก ก. : ก {a i }=a 1,,a n a i N,

Big-O Notation and Complexity Analysis

Lecture 1: Asymptotics, Recurrences, Elementary Sorting

CSED233: Data Structures (2017F) Lecture4: Analysis of Algorithms

What is Performance Analysis?

3.1 Asymptotic notation

Asymptotic Analysis. Thomas A. Anastasio. January 7, 2004

Analysis of Algorithms

Analysis of Algorithms

CIS 121 Data Structures and Algorithms with Java Spring Big-Oh Notation Monday, January 22/Tuesday, January 23

Asymptotic Analysis 1

COMP Analysis of Algorithms & Data Structures

COMP Analysis of Algorithms & Data Structures

Data Structures and Algorithms. Asymptotic notation

Lecture 10: Big-Oh. Doina Precup With many thanks to Prakash Panagaden and Mathieu Blanchette. January 27, 2014

Enumerate all possible assignments and take the An algorithm is a well-defined computational

Mat Week 6. Fall Mat Week 6. Algorithms. Properties. Examples. Searching. Sorting. Time Complexity. Example. Properties.

Student Responsibilities Week 6. Mat Properties of Algorithms. 3.1 Algorithms. Finding the Maximum Value in a Finite Sequence Pseudocode

Asymptotic Algorithm Analysis & Sorting

Measuring Goodness of an Algorithm. Asymptotic Analysis of Algorithms. Measuring Efficiency of an Algorithm. Algorithm and Data Structure

Topic 17. Analysis of Algorithms

CS473 - Algorithms I

with the size of the input in the limit, as the size of the misused.

Chapter 0. Prologue. Algorithms (I) Johann Gutenberg. Two ideas changed the world. Decimal system. Al Khwarizmi

The Time Complexity of an Algorithm

Lecture 1: Asymptotic Complexity. 1 These slides include material originally prepared by Dr.Ron Cytron, Dr. Jeremy Buhler, and Dr. Steve Cole.

Algorithms (I) Yijia Chen Shanghai Jiaotong University

CSE 101. Algorithm Design and Analysis Miles Jones Office 4208 CSE Building Lecture 1: Introduction

CS 4407 Algorithms Lecture 2: Growth Functions

Taking Stock. IE170: Algorithms in Systems Engineering: Lecture 3. Θ Notation. Comparing Algorithms

The Time Complexity of an Algorithm

Lecture 3: Big-O and Big-Θ

Big O Notation. P. Danziger

Cpt S 223. School of EECS, WSU

Data Structures and Algorithms CSE 465

CSE 421: Intro Algorithms. 2: Analysis. Winter 2012 Larry Ruzzo

Big O Notation. P. Danziger

CS 161 Summer 2009 Homework #2 Sample Solutions

COMP 9024, Class notes, 11s2, Class 1

Introduction to Algorithms and Asymptotic analysis

Comparison of functions. Comparison of functions. Proof for reflexivity. Proof for transitivity. Reflexivity:

CS Non-recursive and Recursive Algorithm Analysis

Data structures Exercise 1 solution. Question 1. Let s start by writing all the functions in big O notation:

CS 310 Advanced Data Structures and Algorithms

Introduction to Divide and Conquer

P, NP, NP-Complete, and NPhard

CSE 417: Algorithms and Computational Complexity

Lecture 3. Big-O notation, more recurrences!!

Recommended readings: Description of Quicksort in my notes, Ch7 of your CLRS text.

Computer Science 431 Algorithms The College of Saint Rose Spring Topic Notes: Analysis Fundamentals

Fall 2016 Test 1 with Solutions

CS 4407 Algorithms Lecture 3: Iterative and Divide and Conquer Algorithms

data structures and algorithms lecture 2

Divide and Conquer CPE 349. Theresa Migler-VonDollen

Analysis of Algorithms I: Asymptotic Notation, Induction, and MergeSort

Asymptotic Analysis Cont'd

CIS 121. Analysis of Algorithms & Computational Complexity. Slides based on materials provided by Mary Wootters (Stanford University)

2.2 Asymptotic Order of Growth. definitions and notation (2.2) examples (2.4) properties (2.2)

Foundations II: Data Structures and Algorithms

Solutions. Problem 1: Suppose a polynomial in n of degree d has the form

Review Of Topics. Review: Induction

Computational Complexity - Pseudocode and Recursions

CS Data Structures and Algorithm Analysis

Math 391: Midterm 1.0 Spring 2016

Asymptotic Growth Rates and the Big-O Notation

Data Structures and Algorithms

Introduction to Algorithms: Asymptotic Notation

Transcription:

Design and Analysis of Algorithms Part 1 Program Costs and Asymptotic Notations Tom Melham Hilary Term 2015 DAA 2015 1. Program costs and asymptotic notations 1 / 35

Fast computers vs efficient algorithms Web-based innovations (e.g. Google algorithms, wikis, blogs, etc.) rely on fast computers, and efficient algorithms. Which is more important? DAA 2015 1. Program costs and asymptotic notations 2 / 35

Fast computers vs efficient algorithms Cost of running an algorithm is usually a functiont(n) of the input sizen. AssumeT(n) is the number of steps an algorithm takes to compute a size-n problem, and a computer capable of10 10 steps per second. CostT(n) Maximum problem size solvable in (Complexity) 1 second 1 hour 1 year n 10 10 3.6 10 13 3 10 17 n 2 10 5 6 10 6 5 10 8 n 3 2154 33000 680000 2 n 33 45 58 Replacing a cubic by a quadratic algorithm, we can solve a problem roughly 180 times larger (in an hour). Replacing a quadratic by a linear algorithm, we can solve a problem roughly 6000000 times larger! DAA 2015 1. Program costs and asymptotic notations 3 / 35

Why is efficient algorithm design important? Suppose a faster computer is capable of10 16 steps per second. CostT(n) Max. size before Max. size now n s 1 10 6 s 1 n 2 s 2 1000 s 2 n 3 s 3 100 s 3 2 n s 4 s 4 +20 A 10 6 increase in speed results in only a factor-of-100 improvement if complexity isn 3, and only an additive increase of 20 if complexity is 2 n! Conclusions As computer speeds increase... 1.... it is algorithmic efficiency that really determines the increase in problem size that can be achieved. 2.... so does the size of problems we wish to solve. Thus designing efficient algorithms become even more important! DAA 2015 1. Program costs and asymptotic notations 4 / 35

Notations can make all the difference Consider the Roman numerals: I = 1,V = 5,X = 10,L = 50,C = 100,D = 500,M = 1000. How would you compute MDCXLVIII +DCCCXII? Invented in India around AD 600, the decimal system was a revolution in quantitative reasoning. Using ten symbols, combined with a place-value system each symbol has ten times the weight of the one to its right, even large numbers can be written compactly. Arabic mathematicians developed arithmetic methods using the Indian decimals (including decimal fractions), leading to the decimal point. DAA 2015 1. Program costs and asymptotic notations 5 / 35

From Algorism to Algorithms A 9th-century Arabic textbook by the Persian Al Khwarizmi was the key to the spread of the Indian-Arabic decimal arithmetic. He gave methods for basic arithmetics (adding, multiplying and dividing numbers), even the calculation of square roots and digits ofπ. Derived from Al Khwarizmi, algorism means rules for performing arithmetic computations using the Indian-Arabic decimal system. The word algorism devolved into algorithm, with a generalisation of the meaning to Algorithm: a finite set of well-defined instructions for accomplishing some task. DAA 2015 1. Program costs and asymptotic notations 6 / 35

Evaluating algorithms Two questions we ask about an algorithm 1. Is it correct? 2. Is it efficient? Correctness - of utmost importance. It is easy to design a highly efficient but incorrect algorithm. Efficiency with respect to: Running time Space = amount of memory used Network traffic Others. E.g. number of times secondary storage is accessed. DAA 2015 1. Program costs and asymptotic notations 7 / 35

Measuring running time The running time of a program depends on many factors: 1. The running time of the algorithm. 2. The input of the program. Normally the larger the input, the longer it takes to compute. 3. The quality of the implementation (e.g. quality of the code generated by the compiler). 4. The machine running the program. We are concerned with 1. DAA 2015 1. Program costs and asymptotic notations 8 / 35

Sorting The Sorting Problem Input: A sequence ofnnumbersa 0,,a n 1. Output: A permutation a σ(0),,a σ(n 1) of the input such that a σ(0) a σ(1) a σ(n 1). The sequences are typically stored in arrays. The numbers are sometimes called keys. DAA 2015 1. Program costs and asymptotic notations 9 / 35

Insertion sort Insertion sort: Informal description The sorted array (output) is built one entry at a time. At each iteration, the first remaining entry of the input is removed and inserted into the result at the right position, extending the result. The result array after k iterations contains the firstk entries of the input array and is sorted. This goes on until no element is left in the input. DAA 2015 1. Program costs and asymptotic notations 10 / 35

Example j 0 1 2 3 4 5 5 2 4 6 1 3 j 0 1 2 3 4 5 2 5 4 6 1 3 0 1 2 3 4 5 2 4 5 6 1 3 j 0 1 2 3 4 5 2 4 5 6 1 3 0 1 2 3 4 5 1 2 4 5 6 3 0 1 2 3 4 5 6 1 2 3 4 5 6 j j j Observation. At the start of each iteration of the outer for loop, the subarray A[0.. j) consists of the elements originally in A[0.. j) but in sorted order. DAA 2015 1. Program costs and asymptotic notations 11 / 35

Expressing algorithms in pseudocode of [CLRS] Similar to Pascal, C and Java. Pseudocode is for communicating algorithms to humans: many programming issues (e.g. data abstraction, modularity, error handling, etc.) are ignored. English statements are sometimes used. // indicates that the reminder of the line is a comment. (In 2nd edition is used.) Variables are local to the block, unless otherwise specified. Block structure is indicated by indentation. Assignment x = y makes x reference the same object asy. (In 2nd edition is used.) Boolean operators and and or are short-circuiting. DAA 2015 1. Program costs and asymptotic notations 12 / 35

Insertion sort in pseudocode INSERTION-SORT(A) Input: An integer array A Output: Array A sorted in non-decreasing order 1 forj = 1 to A.length 1 2 key = A[j] 3 // InsertA[j] into the sorted sequence A[0..j). 4 i = j 5 whilei > 0 anda[i 1] > key 6 A[i] = A[i 1] 7 i = i 1 8 A[i] = key DAA 2015 1. Program costs and asymptotic notations 13 / 35

Loop-invariant approach to correctness proof Three key components of a loop-invariant argument: 1. Initialization: Prove that invariant (I) holds prior to first iteration. 2. Maintenance: Prove that if (I) holds just before an iteration, then it holds just before the next iteration. 3. Termination: Prove that when the loop terminates, the invariant (I), and the reason that the loop terminates, imply the correctness of the loop-construct The approach is reminiscent of mathematical induction: 1. Initialization corresponds to establishing the base case. 2. Maintenance corresponds to establishing the inductive case. 3. The difference is that we expect to exit the loop, but mathematical induction establishes a result for all natural numbers. DAA 2015 1. Program costs and asymptotic notations 14 / 35

Correctness of INSERTION-SORT Invariant of outer loop: 1 j n and the subarray A[0.. j) consists of the elements originally in A[0..j) but in sorted order. (n = A.length) Initialization: When j = 1, then subarray A[0..j) is a singleton and therefore sorted. Termination. The outer for loop terminates when j = n. Thus, with j = n, the invariant reads: A[0..n) consists of the elements originally in A[0.. n) but in sorted order. Maintenance. Suppose input is sequencea 0,,a n 1. We aim to prove: Supposej = k. If at the start of an iteration 1 k n anda[0..j) consists ofa 0,,a j 1 in sorted order, then at the start of the next iteration (i.e. j = k+1), 1 k+1 n anda[0..j) consists of a 0,,a j 1 in sorted order. DAA 2015 1. Program costs and asymptotic notations 15 / 35

Correctness of INSERTION-SORT, contd. Body of inner while loop works by moving each ofa[j 1],A[j 2],, by one position to the right until the proper position for key is found, and where it is then inserted. Suppose just before reaching the inner loopa[0..j) contains the sorted sequencea 0, a j 1. Invariant of inner loop: 0 i j and A[0..i)A[i+1..j+1) contains the sequencea 0, a j 1 and all elements ina[i+1..j+1) are bigger than key. Initialization. Fori = j, invariant is true as nothing has moved yet. Termination. Eitheri = 0 ora[i 1] key, and all elements in A[i+1..j+1) are bigger than key. Hence A[0..i)keyA[i+1..j+1) is the same sequence asa 0, a j 1 with key inserted in the correct position. Maintenance. Using the fact that the body is only executed ifi > 0 and A[i 1] > key the invariant is true again after executing both statements of the body. DAA 2015 1. Program costs and asymptotic notations 16 / 35

Running time analysis Assume line i takes time c i, a constant. Forj = 1,,n 1, let t j be the number of times the test of the while loop is excecuted for that value ofj. When a for or while loop exits normally, the test is executed one more time than the loop body. Running time is (cost of statement) (number of time statement executed) all statements LetT(n) be the running time of INSERTION-SORT wheren = A.length. We have: T(n) = c 1 n+c 2 (n 1)+c 3 (n 1)+c 4 (n 1)+c 5 n 1 j=1 t j +c 6 n 1 j=1 (t j 1)+c 7 n 1 j=1 (t j 1)+c 8 (n 1) DAA 2015 1. Program costs and asymptotic notations 17 / 35

Worst-case analysis The input array contains distinct elements in reverse sorted order i.e. is strictly decreasing. Why? Because we have to compare key = a j with every element to left of the(j +1)th element, and so, compare with j elements in total. Thust j = j +1. We have n 1 j=1 t j = n 1 n(n+1) j=1 (j +1) = 1, and 2 so, T(n) = c 1 n+c 2 (n 1)+c 4 (n 1)+c 5 ( n(n+1) 2 1) +c 6 ( n(n 1) 2 )+c 7 n(n 1) 2 +c 8 (n 1) = an 2 +bn+c for appropriate a,b and c. Hence T(n) is a quadratic function ofn. DAA 2015 1. Program costs and asymptotic notations 18 / 35

Best-case analysis The array is already sorted. Always finda[i] key upon the first iteration of the while loop (when i = j). Thust j = 1. T(n) = c 1 n+c 2 (n 1)+c 4 (n 1)+c 5 (n 1)+c 8 (n 1) = (c 1 +c 2 +c 4 +c 5 +c 8 )n+(c 2 +c 4 +c 5 +c 8 ) I.e. T(n) is a linear function ofn. DAA 2015 1. Program costs and asymptotic notations 19 / 35

Average-case analysis Randomly choosennumbers as input. On average key ina[j] is less than half the elements in A[0..j) and greater than the other half, and so, on average while loop has to look halfway through the sorted subarray A[0..j) to decide where to drop key. Hence t j = j/2. Although average-case running time is approximately half that of the worst-case, it is still quadratic. Moral. Average-case complexity can be asymptotically as bad as the worst! Average-case analysis is not straightforward: What is meant by average input depends on the application. The mathematics can be difficult. DAA 2015 1. Program costs and asymptotic notations 20 / 35

Features of insertion sort: A summary Minuses Worst-case quadratic time. Pluses Efficient on small data sets. Efficient (linear-time) on already sorted (or nearly sorted) input. Stable: Relative order of elements with equal keys is maintained. In-place: Only a constant amount of extra memory space (other than that which holds the input) is required, regardless of the size of the input. Online: it can sort a list as it is received. DAA 2015 1. Program costs and asymptotic notations 21 / 35

Insertion sort in Haskell > isort = foldr insert [] > insert :: Ord a => a -> [a] -> [a] > insert x [] = [x] > insert x (y:ys) = > if x > y then y : insert x ys > else x : y : ys Worst case analysis: T insert (m,0) c 1 and T insert (m,n+1) T insert (m,n)+c 2 gives T insert (m,n) c 2 n+c 1 T isort (0) c 3 and T isort (n+1) T insert (m,n)+t isort (n)+c 4 gives T isort (n) c 2 ( n(n 1) 2 )+(c 1 +c 4 )n+c 3. DAA 2015 1. Program costs and asymptotic notations 22 / 35

The Big-O Notation Letf,g : N R + be functions. Define the set O(g(n)) := {f : N R + : n 0 N +. c R +. n. n n 0 f(n) c g(n)} In words,f O(g) just if there exist a positive integer n 0 and a positive real c such that for all n n 0, we have f(n) c g(n). Informally O(g) is the set of functions that are bounded above byg, ignoring constant factors, and ignoring a finite number of exceptions. Equivalently, f O(g) just if there exist positive reals a and b such that for all n, we have f(n) a g(n)+b. (Proof: exercise) DAA 2015 1. Program costs and asymptotic notations 23 / 35

Examples O(g(n)) := {f : N R + : n 0 N +. c R +. n. n n 0 f(n) c g(n)} 1. 3 98 O(1). Just taken 0 = 1 and c = 3 98. 2. 5n 2 +9 O(n 2 ). Taken 0 = 3 and c = 6. Then for for all n n 0, we have9 n 2, and so5n 2 +9 5n 2 +n 2 = 6n 2 = cn 2. 3. Takeg(n) = n 2 and f(n) = 7n 2 +3n+11. Thenf O(g). 4. Some more functions in O(n 2 ): 1000n 2,n, n 1.9999, n 2 /lglglgn and 6. DAA 2015 1. Program costs and asymptotic notations 24 / 35

A shorthand for Big O We say g is an asymptotic upper bound forf just if f O(g). Instead of writing f O(g) we usually write (read f is Big-O ofg ). Informally f(n) = O(g(n)) means f(n) = O(g(n)) f(n) grows no faster than (a constant factor of)g(n), for sufficiently large n. DAA 2015 1. Program costs and asymptotic notations 25 / 35

Big-O on the RHS of an equation It is convenient to write f 1 (n) = f 2 (n)+o(g(n)) instead off 1 (n) f 2 (n) = O(g(n)) (which is the proper way). DAA 2015 1. Program costs and asymptotic notations 26 / 35

Pitfalls of Big O notation When writing f(n) = O(g(n)) bear in mind that it is a shorthand forf(n) O(g(n)). So = does not have all the usual properties of equality. In particular it is not symmetric! Examples n = O(n 2 ) butn 2 O(n). n = O(n 3 ) andn 2 = O(n 3 ) butn n 2. DAA 2015 1. Program costs and asymptotic notations 27 / 35

So why use the big-o notation? More than 100 years old (so very standard mathematical notation), it is too well established to change now. We can and should read = as is. Note that is does not imply equality. We already abuse the = symbol in computer science... DAA 2015 1. Program costs and asymptotic notations 28 / 35

Properties of Big-O Notation. log 2 n is sometimes written lgn; and log e n is sometimes written lnn ( n for natural logarithm). DAA 2015 1. Program costs and asymptotic notations 29 / 35

Properties of Big-O Lemma 1. Let f,g,h : N R +. Then: 1. For every constantc > 0, if f O(g) then cf O(g). 2. For every constant c > 0, if f O(g) then f O(c g). (E.g.f(n) = O(lnn) iff f(n) = O(log 10 n). Thus it is preferable to be neutral and write f(n) = O(logn).) 3. Iff 1 O(g 1 ) andf 2 O(g 2 ) then f 1 +f 2 O(g 1 +g 2 ). 4. Iff 1 O(g 1 ) andf 2 O(g 2 ) then f 1 +f 2 O(max(g 1,g 2 )). 5. Iff 1 O(g 1 ) andf 2 O(g 2 ) then f 1 f 2 O(g 1 g 2 ). 6. Iff O(g) andg O(h) then f O(h). 7. Every polynomial of degreel 0 is in O(n l ). 8. For anyc > 0 in R, we havelg(n c ) O(lg(n)). 9. For every constantc,d > 0, we havelg c (n) O(n d ). 10. For every constantc > 0 andd > 1, we haven c O(d n ). DAA 2015 1. Program costs and asymptotic notations 30 / 35

Question and Example We have seen that O(lgn) = O(lnn). Question. IsO(2 lgn ) the same aso(2 lnn )? No. O(2 lnn ) = O(n 0.6931... ). Recall: a log b c = c log b a. DAA 2015 1. Program costs and asymptotic notations 31 / 35

Question and Example Example. We show 57n 3 +4n 2 lg 5 (n)+17n+498 O(n 3 ) by appealing to Lemma 1: lg 5 (n) O(n) 9 4n 2 lg 5 (n) O(4n 3 ) 5 57n 3 +4n 2 lg 5 (n)+17n+498 O(57n 3 +4n 3 +17n+498) 3 57n 3 +4n 3 +17n+498 O(n 3 ) 7 57n 3 +4n 2 lg 5 (n)+17n+498 O(n 3 ) 6 DAA 2015 1. Program costs and asymptotic notations 32 / 35

Big-Θ We write f = Θ(g), read g is an asymptotic tight bound off, to mean f = O(g) andg = O(f). I.e. f = Θ(g) just if there are positive reals c 1 and c 2 and a positive numbern 0 such that for all n n 0, we have c 1 g(n) f(n) c 2 g(n). We think off and g as having the same order of magnitude. Examples 1. 5n 3 +88n = Θ(n 3 ) 2. 2+sin(logn) = Θ(1) 3. n! = Θ(n n+1/2 e n ). (A consequence of Stirling s Approximation.) DAA 2015 1. Program costs and asymptotic notations 33 / 35

Big-Ω The Big-O notation is useful for upper bounds. There is an analogous notation for lower bounds. We write f = Ω(g) to mean there are a positive numbern 0 and a positive real c such that for all n n 0, we havef(n) cg(n). We say g is an asymptotic low bound off just iff Ω(g). Example 1. n n = Ω(n!) 2. exp(n) = Ω(n 10 ). Exercise Prove thatf = Ω(g) iff g = O(f). DAA 2015 1. Program costs and asymptotic notations 34 / 35

Some useful results Logarithms. Recall the following useful facts. Leta,b,c > 0. a = b log b a log b a = log c a log c b log b a = 1 log a b a log b c = c log b a A form of Stirling s approximation: n! = 2πn ( n e) n ( 1+O ( )) 1 n DAA 2015 1. Program costs and asymptotic notations 35 / 35