# Random Forest

## Random Forest :

Random Forest is a machine learning algorithm that belongs to the ensemble learning method. It is used for classification and regression problems. It creates a forest of decision trees, where each tree is trained on a different set of data and makes a prediction. The final prediction is made by taking the average or majority vote of all the trees. This process helps in reducing the variance and overfitting of a single decision tree, resulting in a more robust model.
Example 1:
Suppose we want to predict whether a person is likely to have a heart disease or not. For this, we have a dataset of various features such as age, blood pressure, cholesterol level, and so on. To predict the outcome, we can use a Random Forest algorithm.
Initially, the algorithm will randomly select a subset of the data and train a decision tree on it. Then, it will select another subset of the data and train another decision tree. This process is repeated until all the subsets are used to train the decision trees.
Now, suppose we have trained 10 decision trees on different subsets of the data. When a new test sample is given to the model, each tree will make a prediction on whether the person has a heart disease or not. If six out of the ten trees predict that the person has a heart disease, then the final prediction will be that the person has a heart disease.
Example 2:
Consider a problem of predicting the salary of an employee based on his/her years of experience and education level. We can use a Random Forest algorithm to solve this problem.
The algorithm will again randomly select a subset of the data and train a decision tree on it. Then, it will select another subset of the data and train another decision tree. This process is repeated until all the subsets are used to train the decision trees.
Let’s say we have trained 10 decision trees on different subsets of the data. When a new test sample is given to the model, each tree will make a prediction on the salary of the employee. If six out of the ten trees predict a salary of \$50,000, then the final prediction will be \$50,000.
In both examples, the Random Forest algorithm has helped in reducing the variance and overfitting of a single decision tree. It has also provided a more robust and accurate prediction by taking the average or majority vote of all the trees.