October 19, 2024

XGBoost

XGBoost is based on gradient boosting trees, it excels in computational efficiency, flexibility, and features. It is a model commonly used in machine learning competitions.

https://arxiv.org/abs/1603.02754

Loss Function and Regularization

The T represents the number of terminal nodes. The variable w is a vector that contains the values of all leaf nodes, which are the average values of y for the samples in each leaf node. γ and λ are hyperparameters.

In other words, XGBoost is a modification of gradient boosting trees, optimized by adding a regularization term to the original loss function.