Big-O Notation
Last updated
Last updated
As computer scientists started to develop various algorithms for a solutions, they needed a way to classify the effectiveness of their algorithms, and they also required a way to prove that a new algorithm is better than the old by mathematical proof & analysis.
Due to this, three notations methods were developed:
Big - O Notation (the worst case scenario)
Big - Theta Notation (the average case scenario)
Big- Omega Notation (the best case scenario)
When we analyze an algorithm we can look at all three scenarios and see their effectiveness. The main notation that we will focus will be the Big-O Notation. We focus on the Big-O because it shows us that at worst our algorithm will perform at X, and if X is better than our previous algorithm, it is definitely better.
Big O Notation: a mathematical notation that describes the limiting behaviour of a function as its input/parameter/argument approaches infinity.
Consider the Big-O to tell us the āWorst case scenario performanceā of our algorithm
We use Big-O Notation for:
Algorithm Proof: To prove that our algorithm A is better than algorithm B, we must have a proof that our Big-O notation is better
Measure Performance, Run-time, or Disk Usage: Our algorithms are designed to solve problems; however, reaching the solutions must not be feasible due to time or disk-space constraints
Mathematically Formalizing our Algorithms: Different hardware will output different runtimes; therefore, we needed a formal mathematical analysis
Big-O is Classified with a Notation of:
The following is a common list of complexities from best to worst performance:
Time & Space Complexity
Time: the measure of how long an algorithm takes
Space: the measure of how much disk space that the algorithm requires from the computer
Index
l[i]
O(1)
Store
l[i] = 0
O(1)
Length
len(l)
O(1)
Append
l.append(5)
O(1)
Pop
l.pop()
O(1)
same as l.pop(-1), popping at end
Clear
l.clear()
O(1)
similar to l = []
Slice
l[a:b]
O(b-a)
l[1:5]:O(l)/l[:]:O(len(l)-0)=O(N)
Extend
l.extend(ā¦)
O(len(ā¦))
depends only on len of extension
Construction
list(ā¦)
O(len(ā¦))
depends on length of argument
āā
āā
āā
āā
check ==, !=
l1 == l2
O(N)
Insert
l[a:b] = ā¦
O(N)
Delete
del l[i]
O(N)
Remove
l.remove(ā¦)
O(N)
Containment
x in / not in l
O(N)
searches list
Copy
l.copy()
O(N)
Same as l[:] which is O(N)
Pop
l.pop(0)
O(N)
Extreme value
min(l)/max(l)
O(N)
Reverse
l.reverse()
O(N)
Iteration
for v in l:
O(N)
āā
āā
āā
āā
Sort
l.sort()
O(N Log N)
key/reverse doesnāt change this
Multiply
k*l
O(k N)
5*l is O(N): len(l)*l is O(N**2)
Tuples support all operations that do not mutate the data structure (and with the same complexity classes).
Length
len(s)
O(1)
Add
s.add(5)
O(1)
Containment
x in/not in s
O(1)
compare to list/tuple - O(N)
Remove
s.remove(5)
O(1)
compare to list/tuple - O(N)
Discard
s.discard(5)
O(1)
Pop
s.pop()
O(1)
compare to list - O(N)
Clear
s.clear()
O(1)
similar to s = set()
āā
āā
āā
āā
Construction
set(ā¦)
len(ā¦)
check ==, !=
s != t
O(min(len(s),lent(t))
<=/<
s <= t
O(len(s1))
issubset
>=/>
s >= t
O(len(s2))
issuperset s <= t == t >= s
Union
s
t
O(len(s)+len(t))
Intersection
s & t
O(min(len(s),lent(t))
Difference
s - t
O(len(t))
Symmetric Difference
s ^ t
O(len(s))
āā
āā
āā
āā
Iteration
for v in s:
O(N)
Copy
s.copy()
O(N)
Index
d[k]
O(1)
Store
d[k] = v
O(1)
Length
len(d)
O(1)
Delete
del d[k]
O(1)
get/setdefault
d.method
O(1)
Pop
d.pop(k)
O(1)
Pop item
d.popitem()
O(1)
Clear
d.clear()
O(1)
similar to s = {} or = dict()
Views
d.keys()
O(1)
āā
āā
āā
āā
Construction
dict(ā¦)
len(ā¦)
āā
āā
āā
āā
Iteration
for k in d:
O(N)
all forms: keys, values, items
We look at every value of N from 1 to N; therefore, this is O(n).
Minor computations such as: += 1
and n % divisor
does not affect the overall classification
The refactored version only analyzes value from 1 to SquareRoot(n); therefore, the refactored version is: