Want to share your content on python-bloggers? click here.
We already know a single decision tree can work surprisingly well. The idea of constructing a forest from individual trees seems like the natural next step.
Today you’ll learn how the Random Forest classifier works and implement it from scratch in Python. This is the sixth of many upcoming from-scratch articles, so stay tuned to the blog if you want to learn more. The links to the previous articles are located at the end of this piece.
The article is structured as follows:
- Introduction to Random Forest
- Math Behind Random Forest
- From-Scratch Implementation
- Model Evaluation
- Comparison with Scikit-Learn
- Conclusion
You can download the corresponding notebook here.
Introduction to Random Forest
Just like decision trees, random forests are a non-parametric model used for both regression and classification tasks. If you understood the previous article on decision trees, you’ll have no issues understanding this one.
Needless to say, but that article is also a prerequisite for this one, for obvious reasons.
The entire random forest algorithm is built on top of weak learners (decision trees), giving you the analogy of using trees to make a forest. The term “random” indicates that each decision tree is built with a random subset of data.
Here’s an excellent image comparing decision trees and random forests:
As simple as that.
The random forest algorithm is based on the bagging method. It represents a concept of combining learning models to increase performance (higher accuracy or some other metric).
In a nutshell:
- N subsets are made from the original datasets
- N decision trees are build from the subsets
- A prediction is made with every trained tree, and a final prediction is returned as a majority vote
Here’s a diagram to drive these points home:
Let’s go over the math behind the algorithm next.
Math Behind Random Forest
Good news – no math today!
The math behind random forests is the same as with decision trees. You only need to implement two formulas – entropy and information gain.
If these sound like a foreign language, please refer to the previous article. Both concepts are discussed in great detail there.
The rest of the article assumes you’re familiar with the inner workings of decision trees, as it is required to build the algorithm from scratch.
From-Scratch Implementation
We’ll need three classes this time:
Node
– implements a single node of a decision treeDecisionTree
– implements a single decision treeRandomForest
– implements our ensemble algorithm
The first two classes are identical as they were in the previous article, so feel free to skip ahead if you already have them written.
Let’s start with the Node
class. It is here to store the data about the feature, threshold, data going left and right, information gain, and the leaf node value. All are initially set to None
. The root and decision nodes will contain values for everything besides the leaf node value, and the leaf node will contain the opposite.
Here’s the code for the class (alongside the library imports):
Let’s implement the decision tree classifier next. It will contain a bunch of methods, all of which are discussed below:
__init__()
– the constructor, holds values formin_samples_split
andmax_depth
. These are hyperparameters. The first one is used to specify a minimum number of samples required to split a node, and the second one specifies a maximum depth of a tree. Both are used in recursive functions as exit conditions_entropy(s)
– calculates the impurity of an input vectors
_information_gain(parent, left_child, right_child)
calculates the information gain value of a split between a parent and two children_best_split(X, y)
function calculates the best splitting parameters for input featuresX
and a target variabley
. It does so by iterating over every column inX
and every threshold value in every column to find the optimal split using information gain_build(X, y, depth)
function recursively builds a decision tree until stopping criteria is met (hyperparameters in the constructor)fit(X, y)
function calls the_build()
function and stores the built tree to the constructor_predict(x)
function traverses the tree to classify a single instancepredict(X)
function applies the_predict()
function to every instance in matrixX
.
Yes, it’s a lot, but you should already feel comfortable with this. Here’s the code snippet for a single decision tree:
Finally, let’s build the forest. This class is built on top of a single decision tree and has the following methods:
__init__()
– the constructor, holds hyperparameter values for the number of trees in the forest, minimum samples split, and maximum depth. It will also hold individually trained decision trees once the model is trained_sample(X, y)
function applies bootstrap sampling to input features and input targetfit(X, y)
function trains the classifier modelpredict(X)
function makes predictions with individual decision trees and then applies majority voting for the final prediction
Code-wise it’s a much simpler class than a decision tree. Here’s the entire snippet:
You might not understand everything fully in one sitting, but this won’t be too much of a challenge if you understood decision trees.
Let’s train and evaluate our model next.
Model Evaluation
Let’s test our classifier next. We’ll use the Iris dataset from Scikit-Learn. The following code snippet loads the dataset and separates it into features (X
) and the target (y
):
Let’s split the dataset into training and testing portions next. The following code snippet does just that, in an 80:20 ratio:
And now let’s do the training. The code snippet below trains the model with default hyperparameters and makes predictions on the test set:
Let’s take a look at the generated predictions (preds
):
And now at the actual class labels (y_test
):
As you can see, both are identical, indicating a perfectly accurate classifier. You can further evaluate the performance if you want. The code below prints the accuracy score on the test set:
If you were to run the above code, the value of 1.0
would get printed, indicating a perfect classifier. The Iris dataset is incredibly easy to classify correctly, so don’t let this fool you.
Let’s compare our classifier to the one built into Scikit-Learn next.
Comparison with Scikit-Learn
We want to know if our model is any good, so let’s compare it with something we know works well — a RandomForestClassifier
class from Scikit-Learn.
You can use the following snippet to import the model class, train the model, make predictions, and print the accuracy score:
As you would expect, we get a perfect accuracy score of 1.0.
And that’s all for today. Let’s wrap things up in the next section.
Conclusion
And there you have it – how to build a forest from the trees. It’s easier than you would think, especially if you consider that random forests are among the top-performing machine learning algorithms today.
You now know how to implement the Decision tree classifier algorithm from scratch. Does that mean you should ditch the de facto standard machine learning libraries? No, not at all. Let me elaborate.
Just because you can write something from scratch doesn’t mean you should. Still, knowing every detail of how algorithms work is a valuable skill and can help you stand out from every other fit and predict data scientist.
Thanks for reading, and please stay tuned to the blog if you’re interested in more machine learning from scratch articles.
Learn More
- Master Machine Learning: Simple Linear Regression From Scratch With Python
- Master Machine Learning: Multiple Linear Regression From Scratch With Python
- Master Machine Learning: Logistic Regression From Scratch With Python
- Master Machine Learning: K Nearest Neighbors From Scratch With Python
- Master Machine Learning: Decision Trees From Scratch With Python
Stay connected
- Follow me on Medium for more stories like this
- Sign up for my newsletter
- Connect on LinkedIn
The post Master Machine Learning: Random Forest From Scratch With Python appeared first on Better Data Science.
Want to share your content on python-bloggers? click here.