In this tutorial, you will learn what master theorem is and how it is used for solving recurrence relations.
The master method is a formula for solving recurrence relations of the form:
T(n) = aT(n/b) + f(n), where, n = size of input a = number of subproblems in the recursion n/b = size of each subproblem. All subproblems are assumed to have the same size. f(n) = cost of the work done outside the recursive call, which includes the cost of dividing the problem and cost of merging the solutions Here, a ≥ 1 and b > 1 are constants, and f(n) is an asymptotically positive function.
An asymptotically positive function means that for a sufficiently large value of n, we have f(n) > 0
.
The master theorem is used in calculating the time complexity of recurrence relations (divide and conquer algorithms) in a simple and quick way.
1. Master Theorem
If a ≥ 1
and b > 1
are constants and f(n)
is an asymptotically positive function, then the time complexity of a recursive relation is given by
T(n) = aT(n/b) + f(n) where, T(n) has the following asymptotic bounds: 1. If f(n) = O(nlogb a-ϵ), then T(n) = Θ(nlogb a). 2. If f(n) = Θ(nlogb a), then T(n) = Θ(nlogb a * log n). 3. If f(n) = Ω(nlogb a+ϵ), then T(n) = Θ(f(n)). ϵ > 0 is a constant.
Each of the above conditions can be interpreted as:
- If the cost of solving the sub-problems at each level increases by a certain factor, the value of
f(n)
will become polynomially smaller thannlogb a
. Thus, the time complexity is oppressed by the cost of the last level ie.nlogb a
- If the cost of solving the sub-problem at each level is nearly equal, then the value of
f(n)
will benlogb a
. Thus, the time complexity will bef(n)
times the total number of levels ie.nlogb a * log n
- If the cost of solving the subproblems at each level decreases by a certain factor, the value of
f(n)
will become polynomially larger thannlogb a
. Thus, the time complexity is oppressed by the cost off(n)
.
2. Solved Example of Master Theorem
T(n) = 3T(n/2) + n2 Here, a = 3 n/b = n/2 f(n) = n2 logb a = log2 3 ≈ 1.58 < 2 ie. f(n) < nlogb a+ϵ , where, ϵ is a constant. Case 3 implies here. Thus, T(n) = f(n) = Θ(n2)
3. Master Theorem Limitations
The master theorem cannot be used if:
- T(n) is not monotone. eg.
T(n) = sin n
f(n)
is not a polynomial. eg.f(n) = 2n
- a is not a constant. eg.
a = 2n
a < 1
Related posts:
Dijkstra's Algorithm
Spanning Tree and Minimum Spanning Tree
Tree Traversal - inorder, preorder and postorder
Selection Sort Algorithm
Linked List Operations: Traverse, Insert and Delete
Divide and Conquer Algorithm
Hash Table
Bubble Sort
Stack Data Structure
Asymptotic Analysis: Big-O Notation and More
Bellman Ford's Algorithm
Linear Search
Quicksort Algorithm
Dynamic Programming
Adjacency List
Backtracking Algorithm
Shell Sort Algorithm
Binary Tree
Decrease Key and Delete Node Operations on a Fibonacci Heap
Queue Data Structure
Why Learn Data Structures and Algorithms?
Deque Data Structure
Merge Sort Algorithm
Depth First Search (DFS)
Tree Data Structure
Linked list Data Structure
Binary Search Tree (BST)
Insertion on a B+ Tree
Pick's Theorem - area of lattice polygons
Deletion from a B+ Tree
Insertion Sort Algorithm
Rabin-Karp Algorithm