Programmers must write programs that are correct, efficient, and maintainable
Prior Focus: is the program implemented correctly?
New Focus: does the program have good performance?
Programmers often ask “does your program have good performance?” Here are some common answers!
“Yes, it was fast enough on my laptop”
“I think it is fast enough for most users”
“I don’t know, I never measured its performance”
“It is really fast if you run it on an Apple M3 chip”
“Let’s wait for our user’s to tell us if it is fast enough”
“One function is too slow — and I think we rarely use it”
A two-week road map for exploring Python program performance
Week 1: Empirical Evaluation of Runtime
How do you conduct timing experiments to measure the performance of a real Python programs?
Week 2: Analytical Evaluation of Time Complexity
How do you use an instruction cost model to prove that it belongs to a specific time complexity class?
Ultimate Goal: create a nuanced description of program efficiency that adapts to different inputs and to different computers
Algorithm analysis challenges
Different Inputs
Varying sizes of input data
Different types of data
Degrees of sortedness
Degrees of duplication
Different data distributions
Different Computers
Diverse hardware setups
Varying system performance
Diverse system building
Different operating systems
Variation in system load
Ultimate goal: experimental and analytical evaluation methods that yield actionable insights that support understanding and prediction
Practical methods for measuring the running time of a program
Implement as a Python function with inputs and outputs
Use the time module to measure running time on inputs
Study trade-offs associated with different implementations
Run function with different inputs & execution environments
Implementing a duplicate detector
from typing import Listdef duplicates(input_list: List[int]) ->bool: n =len(input_list)for i inrange(n):for j inrange(n):if i != j and input_list[i] == input_list[j]:returnTruereturnFalseassert(duplicates([1,2,6,3,4,5,6,7,8]))print(duplicates([1,2,6,3,4,5,6,7,8]))assert(not duplicates([1,2,3,4]))print(not duplicates([1,2,3,4]))
True
True
Returns True if there are any duplicates and False otherwise
What is the performance of the duplicates function? Why?
Explore the duplicate detector
Run duplicate with different inputs and observe its output
Experiment with using both the assert and print statements
Performance of a duplicate detector?
from typing import Listimport timefor i inrange(5): n =1000 start = time.time() duplicates(list(range(n))) time_taken = time.time() - startprint("Time taken for n = {}: {:.15f} seconds".format(n, time_taken))
Time taken for n = 1000: 0.038798570632935 seconds
Time taken for n = 1000: 0.038587331771851 seconds
Time taken for n = 1000: 0.038299322128296 seconds
Time taken for n = 1000: 0.038945674896240 seconds
Time taken for n = 1000: 0.040074110031128 seconds
What is the strategy for calculating the elapsed time?
Do the results suggest that this implementation is fast? Why?
General-purpose function timing
from typing import Callableimport timedef timetrials(function: Callable, n: int, trials: int=10) ->None:"""Time the function with n inputs trials times""" totaltime =0for _ inrange(trials): start = time.time() function(list(range(n))) totaltime += time.time() - startprint("average =%10.7f for n = %d"% (totaltime/trials, n))# conduct a doubling experiment for a provided functionfor n in [50, 100, 200, 400, 800, 1600, 3200]: timetrials(duplicates, n)
average = 0.0000776 for n = 50
average = 0.0003226 for n = 100
average = 0.0011619 for n = 200
average = 0.0054209 for n = 400
average = 0.0243838 for n = 800
average = 0.1035131 for n = 1600
average = 0.4343791 for n = 3200
Explanation of timetrial function
Imports:
Callable from typing: Allows functions as arguments
time: Used for timing the function’s execution
Function timetrials:
Parameters:
function: The function to be timed
n: The size of the input list
trials: Number of times to run the function (default is 10)
Timing Loop:
Runs the function trials times
Measures the time taken for each run and calculates the average
Duplicate detecting experiment
Duplicate Function
Accepts an input list
True if duplicate found
False if no duplicates
n is the list’s length
Double nested for loop
Timing Function
Use the time module
Calculate elapsed time
Consistent number format
Provide a function as input
Run a doubling experiment
Key Observation: properly designed experiments yield meaningful insights into the likely worst-case performance of a Python function
Understanding Doubling Experiments
Doubling Experiments:
Measure performance by doubling input size
Observe changes in execution time
Helps to identify the worst-case performance
Key Insights:
Double Time: Execution time doubles when input size doubles
Suggests linear time complexity or \(O(n)\)
Quadruple Time: Time quadruples when input size doubles
Suggests quadratic time complexity or \(O(n^2)\)
Eight-times Time: Time increases eight-fold when input size doubles
Suggests cubic time complexity or \(O(n^3)\)
Can we improve the performance of duplicate detection?
Avoid doing unnecessary computational steps
duplicates compares each pair of elements twice
Improve speed by only letting j range up to i
Improving duplicate’s speed
def duplicates_restrict_range(input_list: List[int]) ->bool: n =len(input_list)for i inrange(1,n):for j inrange(i):if input_list[i] == input_list[j]:returnTruereturnFalsefor n in [50, 100, 200, 400, 800, 1600, 3200]: timetrials(duplicates_restrict_range, n)
average = 0.0000314 for n = 50
average = 0.0001183 for n = 100
average = 0.0004664 for n = 200
average = 0.0019472 for n = 400
average = 0.0090225 for n = 800
average = 0.0395520 for n = 1600
average = 0.1648121 for n = 3200
Did this improve the performance? How can you tell? Why?
Refactoring duplicate again
def duplicates_any(input_list: List[int]) ->bool: n =len(input_list)returnany(input_list[i] == input_list[j]for i inrange(1,n) for j inrange(i))for n in [50, 100, 200, 400, 800, 1600, 3200]: timetrials(duplicates_any, n)
average = 0.0003846 for n = 50
average = 0.0003303 for n = 100
average = 0.0012351 for n = 200
average = 0.0033171 for n = 400
average = 0.0143196 for n = 800
average = 0.0611093 for n = 1600
average = 0.2582761 for n = 3200
Wait, this new implementation does not seem to be faster!
Reducing code size does not always improve performance
More alternatives to duplicates
def duplicates_sort(input_list: List[int]) ->bool: n =len(input_list) input_list.sort()for i inrange(n-1):if input_list[i] == input_list[i+1]:returnTruereturnFalsefor n in [50, 100, 200, 400, 800, 1600, 3200]: timetrials(duplicates_sort, n)
average = 0.0000039 for n = 50
average = 0.0000063 for n = 100
average = 0.0000115 for n = 200
average = 0.0000298 for n = 400
average = 0.0003715 for n = 800
average = 0.0004212 for n = 1600
average = 0.0002771 for n = 3200
Does sort improve the performance of duplicates?
Using both any and sort
def duplicates_sort_any(input_list: List[int]) ->bool: n =len(input_list) input_list.sort()returnany(input_list[i] == input_list[i+1] for i inrange(n-1))for n in [50, 100, 200, 400, 800, 1600, 3200]: timetrials(duplicates_sort_any, n)
average = 0.0000041 for n = 50
average = 0.0000064 for n = 100
average = 0.0000148 for n = 200
average = 0.0000264 for n = 400
average = 0.0000566 for n = 800
average = 0.0001190 for n = 1600
average = 0.0002517 for n = 3200
Does sort improve the performance of duplicates?
Does the use of any improve or hinder performance?
What is the connection between code size and efficiency?
Using set with a for loop
def duplicates_set_loop(input_list: List[int]) ->bool: s =set()for e in input_list:if e in s:returnTrue s.add(e)returnFalsefor n in [50, 100, 200, 400, 800, 1600, 3200]: timetrials(duplicates_set_loop, n)
average = 0.0000031 for n = 50
average = 0.0000049 for n = 100
average = 0.0000087 for n = 200
average = 0.0000245 for n = 400
average = 0.0000546 for n = 800
average = 0.0000832 for n = 1600
average = 0.0001849 for n = 3200
Is it possible to avoid the use of the for loop?
Using set without a for loop
def duplicates_set(input_list: List[int]) ->bool:returnlen(input_list) !=len(set(input_list))for n in [50, 100, 200, 400, 800, 1600, 3200]: timetrials(duplicates_set, n)
average = 0.0000015 for n = 50
average = 0.0000023 for n = 100
average = 0.0000035 for n = 200
average = 0.0000086 for n = 400
average = 0.0000196 for n = 800
average = 0.0000405 for n = 1600
average = 0.0000796 for n = 3200
What is the fastest approach to duplicate detection?
How did you know that this approach was the fastest?