Home

Gradient Boosting Classifier

Gradient Boosting Classifier

What’s a Gradient Boosting Classifier?

Gradient boosting classifier is a set of machine learning algorithms that include several weaker models to combine them into a strong big one with highly predictive output. Models of a kind are popular due to their ability to classify datasets effectively.

Gradient boosting classifier usually uses decision trees in model building. But how are the values obtained, processed, and classified?

Classification is a process, where the machine learning algorithm is given some data and puts it into discrete classes. These classes are unique per each data and are categorized accordingly. For example, in our e-mail box we have such categories as “inbox” and “spam”, and the mail received is classified according to the letter’s contextual features.

Regression is also a machine learning algorithm, which works based on the results obtained by the ML model. In the other words, we obtain a real value that is also a continuous value (weight, pulse). Regression aims at predicting value (age of a person) based on continuous values (weight, height, etc.)

Gradient boost was introduced by Jerome Friedman, who believed that with small steps it is possible to predict better with a dataset that is being tested.

To make out predictions and build a decision tree, we will need to carry out several steps.

Step one – Gathering and Analyzing Our Data

Gradient Boosting Classifier

In the table above we are using the training data that we have gathered from six patients. The data shows patients’ presence of chest pain, their pulse (beats per minute), weight (underweight, normal, and overweight), and a history of heart disease. Our aim here is to understand how gradient boost fits a model to this training data.

Step two – Odds and Probability Calculating

Gradient Boosting Classifier

Using gradient boost for classification we discover the initial prediction for every patient in the log (odds).

To calculate the overall log (odds), let’s differentiate between the patients, who answered “yes” for heart disease and the ones, who answered “no”. Here, we have 4 patients in the training dataset that answered positively, and two patients that answered negatively. So, the log (odds) that patients have heart disease is

Gradient Boosting Classifier

This number is going to be present in the initial leaf of our tree as an initial prediction.

But how can we use initial prediction for classification? The easiest and smartest way to do so is to convert the log (odds) to probability. The trick here is to use the logistic function.

And our probability will look like this:

Gradient Boosting Classifier

With the help of the log (odds) we obtained primarily, the probability of heart disease we get is

Gradient Boosting Classifier

The number 0.5 is considered to be the probability threshold in making a classification decision tree based on it, so every number above it makes a patient prone to heart disease automatically. For more information click on the link to watch ROC and AUC machine learning curves.

Step three – Residual Calculating

We perform residual calculating to get the difference between the observed and the predicted values. We cannot classify every patient in the training dataset as the one that surely has heart disease because two of these patients did not confirm any heart deviations. So, it is best to measure the initial prediction error with the help of getting the pseudo residual number. Let’s take every “yes” answer as 1 and every “no” answer as 0. Get the idea of why we’re doing this from the graph below:

Gradient Boosting Classifier

Here, residual = (binary heart disease – probability) or residual = (yes/no answer – 0.67). We put the obtained results in our table’s new column.

Gradient Boosting Classifier

After calculating the residual for each patient, we’ll obtain new values to work within our decision tree’s leaf of initial prediction.

Gradient Boosting Classifier

Step four – Building a Decision Tree

To build a decision tree we will need to use the chest pain, pulse, and weight data to predict the tree leaves and residuals. Thus, it is necessary to investigate which column will best describe the obtained results. To do this, we are going to divide our training data into three subtables and build three smaller trees – three weaker models to merge into a strong one later.

1. Chest pain (binary dataset)

If the answer is “yes” then we will need to find the Residual Sum of Squares (RSS) and the average value of this positive answer.

Gradient Boosting Classifier

To find the average of the “yes” answer, we should take all the patients, who answered positively, add these numbers and multiply by the quantity of the answers, which is 3. For instance,

Gradient Boosting Classifier

Residual sum of Squares or RSS is the sum of the squares of residuals, which indicates errors predicted from actual values of the data set. Small RSS shows that the model perfectly fits the data. Here, average1 and RSS1 are the obtained results, which correspond to the condition of our training model, while average2 and RSS2 are the ones, which do not.

Gradient Boosting Classifier

The formula above shows Уi as an element from the residual column. And Ӯ as the average number.

Gradient Boosting Classifier

As there is also a “no” answer, we should take it into account and perform the same calculations with regards to the patients, who answered negatively: add the numbers and multiply by 3.

Gradient Boosting Classifier Gradient Boosting Classifier

Based on the average value and RSS calculations, we will obtain the following tree leaves:

Gradient Boosting Classifier

Here, we have two leaves with residuals but if we want to count the data error, we need to add RSS1 and RSS2 and the result will be the following:

Gradient Boosting Classifier

2. Weight (categorical dataset)

To find out the error in the categorical data, it is necessary to divide (or categorize) the weight into such subsections as underweight (lower than normal), normal, and overweight (more than normal).

Gradient Boosting Classifier

To find residuals and the RSS here, we are following the same steps we carried out before.

Gradient Boosting Classifier Gradient Boosting Classifier Gradient Boosting Classifier Gradient Boosting Classifier

3. Pulse (numerical dataset)

To understand the differences and data errors in pulse, we will need to take several pulse indicators of patients. For example, 68, 70, 75, 88, 95, and 115 beats per minute. Pulse is a numerical value, where the condition is variable. So, we take our pulse values and classify them according to the order of growth. Then, we will need a graph to visualize the variables and the obtained residuals.

Gradient Boosting Classifier

We take the first two values of pulse and calculate their average result. E.g. (68+70)/2=69. Then we show this result as a red line on the graph.

Gradient Boosting Classifier

Afterward, we aim to try and find the residual average of the left and the right sides on the graph. As we have only one element on the left, our average is going to be the following:

Gradient Boosting Classifier

As the average result is 0.33 we need to show it on our graph. E.g.

Gradient Boosting Classifier

Further, our calculating shifts to the average of the right side of the graph. And it will be:

Gradient Boosting Classifier

We’re showing this result on the graph too.

Gradient Boosting Classifier

So, the final step is to calculate residuals. This can be done with the help of the following formulas.

Gradient Boosting Classifier Gradient Boosting Classifier

We need to perform the same calculation with all the neighboring values of the pulse. Doing so, we obtain the following results:

Gradient Boosting Classifier

Let’s also calculate the same average1/average2, RSS1/RSS2, and the overall RSS value as in the examples above.

Gradient Boosting Classifier Gradient Boosting Classifier Gradient Boosting Classifier Gradient Boosting Classifier Gradient Boosting Classifier

Gradient Boosting Classifier Gradient Boosting Classifier

After we obtained all the neighboring results it is necessary to select the best minimal option. This result has been achieved between the pulse range between 70 and 75. Based on this minimal number we can build the following tree:

Gradient Boosting Classifier

On building a tree with the residuals, the smallest RSS was when we obtained the Weight= Under normal value. So, we take this value as the root of the tree.

Gradient Boosting Classifier

Then, it is visually shown that we have only one value on the left leaf and five values on the right leaf. So, we need to carry out the same calculations for the right leaf and obtain a new tree.

Gradient Boosting Classifier

After the calculations are done, we input the obtained data to the right leaf of the previous tree and get the following results:

Gradient Boosting Classifier

As we have our data divided into the smallest groups (not more than 3 elements on one leaf) we can move to the next step.

Step 5 – Calculating the Output Value

To calculate the output value we will need to use the following formula:

Gradient Boosting Classifier

This formula is the common transformation method, which allows calculating the output value for every leaf.

Gradient Boosting Classifier

Inputting the already obtained values in the formula we will get the new tree with an output value.

Gradient Boosting Classifier

Step Six – Probability Calculating Based on New Values

This step requires updating the Predictions section with the new data. So, we are combining the initial leaf with the new tree. And this new tree is scaled by a learning rate, which is 0.8 and it is meant only for illustrative purposes.

Gradient Boosting Classifier

This calculation is the same we did before at the beginning of the article. However, the output we get is completely new. And again, after finding out the new probability, let’s find the new residual numbers.

Gradient Boosting Classifier Gradient Boosting Classifier

Having the new residuals data it is possible to build a new tree.

Gradient Boosting Classifier

The process of tree-building repeats until there is a maximum number of trees specified or the residuals become as small as possible.

To make the example simple, a grading boost has been configured to just two versions of trees. Here, there’s a need to classify a new person as someone who has heart disease or doesn’t have this condition. So, we are doing the same prediction, and calculating the potential probability.

Let’s input into the formula our learning rate, which equals 0.8, and log(odds), which is equal to 0.69. Doing so, we will obtain the following:

Gradient Boosting Classifier

But to show you more in detail, imagine we have a new patient and want to calculate the probability of heart disease of this patient.

Gradient Boosting Classifier

Let’s calculate our log(odds) predicted with the formula we have:

Gradient Boosting Classifier

The result will be:

Gradient Boosting Classifier

Using our probability formula we have mentioned in Step 2 we can get our next result.

Gradient Boosting Classifier

So, based on the achieved results, our new patient can have heart disease with the probability of 0.95.

Summing Up

The current overview of gradient boosting classifier is shown on a training dataset, but that is the same way it can be used on the real datasets. For instance, if there is a real need to predict whether the patient has a probability of heart disease at present or in the future or not. Thus, now you have an idea of what a gradient boosting classifier is and how it works in classification and tree-building to get accurate predictions and results.

Tags:

Leave a Reply

Your email address will not be published. Required fields are marked *