Predicting complex skill learning with boosting
We will revisit our Skillcraft dataset in this section--this time in the context of another boosting technique known as stochastic gradient boosting. The main characteristic of this method is that in every iteration of boosting, we compute a gradient in the direction of the errors that are made by the model trained in the current iteration.
This gradient is then used in order to guide the construction of the model that will be added in the next iteration. Stochastic gradient boosting is commonly used with decision trees, and a good implementation in R can be found in the gbm
package, which provides us with the gbm()
function. For regression problems, we need to specify the distribution
parameter to be gaussian
. In addition, we can specify the number of trees we want to build (which is equivalent to the number of iterations of boosting) via the n.trees
parameter, as well as a shrinkage
parameter that is used to control the algorithm's learning...