LogitBoost

From Wikipedia, the free encyclopedia
Jump to navigation Jump to search

In machine learning and computational learning theory, LogitBoost is a boosting algorithm formulated by Jerome Friedman, Trevor Hastie, and Robert Tibshirani.

The original paper casts the AdaBoost algorithm into a statistical framework.[1] Specifically, if one considers AdaBoost as a generalized additive model and then applies the cost function of logistic regression, one can derive the LogitBoost algorithm.[2]

Minimizing the LogitBoost cost function

[edit | edit source]

LogitBoost can be seen as a convex optimization. Specifically, given that we seek an additive model of the form

f=tαtht

the LogitBoost algorithm minimizes the logistic loss:

ilog(1+eyif(xi))

See also

[edit | edit source]

References

[edit | edit source]
  1. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  2. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).