Introduction
Welcome to the first blog in my algorithm design series! 🚀 In this series, I'll be sharing various algorithm design techniques that you should use. You may already know these, but still, I'll recommend you to read these articles once. You could at least give me a view. 😄 Jokes apart, in these articles, everything will be precise and to the point. So you could get the full value from these articles. Now, without further ado, let me give you a brief introduction about what all things I'll be covering in this series. In this article, I'll provide you with a brief introduction to the concepts, and in the upcoming articles, we'll delve deep into these concepts.
The various algorithm design techniques are:
Brute Force
Greedy Algorithms
Divide-and-Conquer, Decrease-and-Conquer
Dynamic Programming
Reduction / Transform-and-Conquer
Backtracking and Branch-and-Bound
Brute Force Algorithm
The brute force algorithm is an approach that comes directly to our minds after viewing a problem. This is usually a straightforward approach based on the problem statement. We could use this approach to solve problems with small-sized datasets. Let me give you an example.
Let's say you want to find the shortest path from your home to your school. The brute force solution in this case would be to consider every possible combination of roads and intersections to reach your destination. However, this method is highly inefficient, especially if there are a lot of roads. I hope now you understand what I'm trying to say. 😅
Some examples of brute force algorithms are:
Bubble-Sort
Selection-Sort
Sequential search in an array
Computing pow(a, n) by multiplying a, n times
Convex hull problem
String matching
Exhaustive search: Travelling salesman, Knapsack, etc.
Greedy Algorithm
Greedy algorithms are used to solve optimization problems. These problems usually involve finding the maximum or minimum of something from the given data. In a greedy algorithm, we find the solution through a sequence of steps. At each step, a choice is made that is locally optimal.
What does "locally optimal" mean?
First, we take a small part of the provided data, and then we find the optimal solution within that data. After that, we'll again take some data and repeat the process. Finding the optimal solution within the data we've considered is called locally optimal.
Some examples of greedy algorithms are:
Minimal spanning tree: Prim's algorithm, Kruskal's algorithm
Dijkstra's algorithm for single-source shortest path problem
Greedy algorithm for the Knapsack problem
The coin exchange problem
Huffman trees for optimal encoding
Divide-and-Conquer, Decrease-and-Conquer
In Divide-and-Conquer algorithms, as the name suggests, we first divide the problem into similar subproblems, and then we conquer each problem one by one. It involves three basic steps:
First, we divide the problem into similar sub-problems.
Then, we solve each one of these subproblems individually.
Finally, we combine the results of the subproblems to produce the desired result.
Divide-and-Conquer and Decrease-and-Conquer are very similar, but they have a slight difference between them. In divide and conquer, the size of the problem is reduced by a factor (half, one third, etc.), while in decrease and conquer, the size of the problem is reduced by a constant.
Some examples of divide-and-conquer algorithms are:
Merge-Sort algorithm (using recursion)
Quick sort algorithm (using recursion)
Computing the length of the longest path in a binary tree (using recursion)
Computing Fibonacci numbers (using recursion)
Quick-hull
Some examples of decrease-and-conquer algorithms are:
Computing pow(a, n) by calculating pow(a, n/2) using recursion.
Binary search in a sorted list (using recursion)
Searching in BST
Insertion-Sort
Graph traversal algorithms (DFS and BFS)
Topological sort
Warshall’s algorithm (using recursion)
Permutations (Minimal change approach, Johnson-Trotter algorithm)
Computing a median, Topological sorting, Fake-coin problem (Ternary search)
Dynamic Programming
In divide and conquer, many times the recursively solved subproblems can result in the same computation being performed multiple times. This problem arises when identical problems occur repeatedly in a recursion.
Dynamic programming is used to avoid this issue by storing the results of subproblems in a table and referring to that table to check if we have already calculated the solution to a subproblem before computing it again.
Dynamic programming is a bottom-up technique in which the smaller subproblems are solved first, and the results of these are used to find the solution for larger subproblems.
Some examples of Dynamic programming are:
Fibonacci numbers computed by iteration.
Warshall’s algorithm for transitive closure implemented by iterations.
Floyd’s algorithm for all-pairs shortest paths.
Reduction / Transform-and-Conquer
These methods work as a two-step process. First, the problem is transformed into a different problem that is already known to us, i.e., a problem for which we already know the optimal solution. Then the problem is solved.
The most common type of transformation is the sorting of an array.
For example, in a given list of numbers, find the two closest numbers. In the brute-force solution, we would find the distance between each element in the array and keep track of the pair with the minimum distance. In this approach, the total time complexity is O(n^2). In the Transform and Conquer solution, we first sort the array in O(n log n) time and then find the closest numbers by scanning the array in another single pass with time complexity O(n). Thus, the total time complexity is O(n log n).
Some examples of
Transform and Conquer are:
Gaussian elimination
Heaps and Heapsort
Backtracking
Have you seen the modern lock with a 3-digit password, where the numbers range from 1 to 9? If you don't have the exact password for the lock, you need to test every combination until you find the right one. You would go from something like "111" to "112" and so on until you reach "999". In this case, what you are doing is called backtracking.
Suppose the lock produces a click sound anytime you come across a correct digit. If we can listen to this sound, it will help you reach the goal much faster. These functions are called Pruning functions or Bounding functions.
Backtracking is a method in which a solution is found by searching through a large but finite number of states. With some pruning or bounding functions, we can narrow down our search.
For all problems (like NP-hard problems) for which there does not exist any other efficient algorithm, we use the backtracking algorithm.
Backtracking problems have the following components:
Initial state
Target/Goal state
Intermediate states
Path from the initial state to the target/goal state
Operators to get from one state to another
Pruning function (optional)
The algorithm starts with the construction of a state tree, whose nodes represent the states. The root node is the initial state, and one or more leaf nodes will be our target state. Each edge of the tree represents some operation. The solution is obtained by searching the tree until a target is found.
Some real-life problems where you could use the backtracking approach are:
Three Monks and Three Demons: There are three monks and three demons on one side of a river. You want to move all of them to the other side using a small boat. The boat can carry only two persons at a time. If, on any shore, the number of demons is more than the number of monks, the demons will eat the monks. How can we move all of these people to the other side of the river safely?
The Farmer's Dilemma: There is a farmer who has a goat, a cabbage, and a wolf. If the farmer leaves the goat with the cabbage, the goat will eat the cabbage. If the farmer leaves the wolf alone with the goat, the wolf will kill the goat. How can the farmer move all his belongings to the other side of the river?
Jug Problem: You are given two jugs, a 4-gallon one and a 3-gallon one. There are no measuring markers on the jugs. A tap can be used to fill the jugs with water. How can you get 2 gallons of water in the 4-gallon jug?
Branch-and-bound
The branch and bound method is used when we can evaluate the cost of visiting each node using a utility function. At each step, we choose the node with the lowest cost to proceed further. Branch-and-bound algorithms are implemented using a priority queue. In branch and bound, we traverse the nodes in a breadth-first manner.
A* Algorithm
A* is an extension of the branch and bound approach. In A*, we select the path with the shortest length from the start to the goal. The total length is estimated as the length traversed so far plus a heuristic estimate of the remaining distance from the goal.
Branch-and-bound will always find an optimal solution, which is the shortest path. A* will also find an optimal solution if the heuristic is accurate. Choosing a good heuristic is the most crucial aspect of the A* algorithm.
Conclusion
Usually, any problem can be solved using a variety of methods. However, we shouldn't always settle for the first method that comes to mind, such as brute force. Some methods yield much more efficient solutions than others. That's why it's necessary to be acquainted with all available methods.
Now, you might wonder how to choose the best method.
First, you should thoroughly comprehend the problem statement.
Second, by being familiar with various problems and their solutions, you can then refine your approach.
Ready to Dive Deeper? Explore the Rest of the Series for In-Depth Algorithm Insights. Don't Miss Out – Subscribe to Our Newsletter for Updates on Future Articles! Subscribe here.
Happy Coding ❤️