A hands-on explanation of Gradient Boosting Regression

Vagif Aliyev
7 min readSep 4, 2020

Introduction

One of the most powerful ways of training models is to train multiple models and aggregate their predictions. This is the main concept of Ensemble Learning. While many flavours of Ensemble Learning exist, some of the most powerful algorithms and Boosting Algorithms. In my previous article, I broke down one of the most popular Boosting Algorithms; Adaptive Boosting. Today, I want to talk about its equally powerful twin; Gradient Boosting.

Boosting & Adaptive Boosting vs Gradient Boosting

Boosting refers to any Ensemble Method that can combine several weak learners(a predictor with poor accuracy) to make a strong learner(a predictor with high accuracy). The idea behind boosting is to train models sequentially, each trying to correct its predecessor.

An Overview Of Adaptive Boosting

In Adaptive Boosting, the main idea occurs with the model assigning a certain weight to each instance, and training a weak learner. Based on the predictor’s performance, it gets assigned its own separate weight based on a weighted error rate. The higher the accuracy of the predictor, the higher its weight, and the more “say” it will have on the final prediction.

--

--

Vagif Aliyev
Vagif Aliyev

Written by Vagif Aliyev

19 y/o student ex-founder of Snapstudy (acquired), founding engineer at Upword.ai