The column measure and Gradient-Free Gradient Boosting

09/24/2019
by   Tino Werner, et al.
0

Sparse model selection by structural risk minimization leads to a set of a few predictors, ideally a subset of the true predictors. This selection clearly depends on the underlying loss function L̃. For linear regression with square loss, the particular (functional) Gradient Boosting variant L_2-Boosting excels for its computational efficiency even for very large predictor sets, while still providing suitable estimation consistency. For more general loss functions, functional gradients are not always easily accessible or, like in the case of continuous ranking, need not even exist. To close this gap, starting from column selection frequencies obtained from L_2-Boosting, we introduce a loss-dependent ”column measure”ν^(L̃) which mathematically describes variable selection. The fact that certain variables relevant for a particular loss L̃ never get selected by L_2-Boosting is reflected by a respective singular part of ν^(L̃) w.r.t. ν^(L_2). With this concept at hand, it amounts to a suitable change of measure (accounting for singular parts) to make L_2-Boosting select variables according to a different loss L̃. As a consequence, this opens the bridge to applications of simulational techniques such as various resampling techniques, or rejection sampling, to achieve this change of measure in an algorithmic way.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/10/2022

Loss-guided Stability Selection

In modern data analysis, sparse model selection becomes inevitable once ...
research
08/19/2019

Gradient Boosting Machine: A Survey

In this survey, we discuss several different types of gradient boosting ...
research
02/06/2020

Robust Boosting for Regression Problems

The gradient boosting algorithm constructs a regression estimator using ...
research
02/27/2017

An update on statistical boosting in biomedicine

Statistical boosting algorithms have triggered a lot of research during ...
research
09/11/2021

Omnipredictors

Loss minimization is a dominant paradigm in machine learning, where a pr...
research
05/04/2018

Valid Inference for L_2-Boosting

We review several recently proposed post-selection inference frameworks ...
research
07/20/2020

Wide Boosting

Gradient boosting (GB) is a popular methodology used to solve prediction...

Please sign up or login with your details

Forgot password? Click here to reset