Boosting is another ensemble method.

Boosting will combine the outcome of many trees to tackle Bias.

In Boosting, trees are grown sequentially. The predictions of Trees are added together. In the case of boosting, smaller trees are preferred.

Method

In boosting, a forest of decision trees are made with the number of splits commonly being . We refer to these trees as “weak learners”.

Each tree is fit to the residuals obtained from the previous tree. With each iteration is there is a learning parameter which is multiplied by the predicted to slow learning.

  1. Set and for all in the set
  2. For do
    1. For a tree with splits and terminal nodes to
    2. We add a shrunken version of the decision tree
    3. Update the residuals
  3. Output the boosted model

Parameters

is the number of fits is the learning rate (0.01, 0.001) is the interaction depth or complexity. For internal nodes there are terminal nodes

Advantages

Improved Accuracy. The Boosting procedure slowly “corrects” what the previous models did not account for.

Reduced Overfitting. The Boosting algorithm starts with “weak learners” (stumps) which learn slowly to the data.

Robustness. Boosting is able to handle noisy data and adapts well to misclassified points.