|
BoostingBoosting is another, very popular style of ensemble classifier. Boosted ensembles of learners are trained in a sequential way, with each new learner focusing on the data that are not yet well classified. The basic outline of a boosted learner is as follows. Take any type of base classifier, but make sure that it can train on weighted data. In other words, rather than attempting to minimize the classification error with our base learner, we will need to minimize a weighted classification error,
Now, beginning with uniform weights Finally, we combine the learners by evaluating them, but rather than a simple average we take a weighted average, multiplying each learner j's decision by a linear coefficient The precise details of this process are typically what distinguishes different boosting algorithms. AdaBoostA very popular algorithm is AdaBoost, which was also the boosting algorithm that first popularized the framework. It is easiest to describe AdaBoost using classes In AdaBoost, we compute the coefficient for each learner j based on its weighted error rate,
The final classifier is given by summing the weighted individual learners and applying a threshold:
Boosting and Surrogate Cost FunctionsBoosting algorithms can be shown to correspond to a particular surrogate loss function (replacement for the classification error rate). In particular, AdaBoost corresponds to the exponential loss function
|