Provably tuning the ElasticNet across instances

07/20/2022
by   Maria-Florina Balcan, et al.
0

An important unresolved challenge in the theory of regularization is to set the regularization coefficients of popular techniques like the ElasticNet with general provable guarantees. We consider the problem of tuning the regularization parameters of Ridge regression, LASSO, and the ElasticNet across multiple problem instances, a setting that encompasses both cross-validation and multi-task hyperparameter optimization. We obtain a novel structural result for the ElasticNet which characterizes the loss as a function of the tuning parameters as a piecewise-rational function with algebraic boundaries. We use this to bound the structural complexity of the regularized loss functions and show generalization guarantees for tuning the ElasticNet regression coefficients in the statistical setting. We also consider the more challenging online learning setting, where we show vanishing average expected regret relative to the optimal parameter pair. We further extend our results to tuning classification algorithms obtained by thresholding regression fits regularized by Ridge, LASSO, or ElasticNet. Our results are the first general learning-theoretic guarantees for this important class of problems that avoid strong assumptions on the data distribution. Furthermore, our guarantees hold for both validation and popular information criterion objectives.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/09/2020

On Coresets For Regularized Regression

We study the effect of norm based regularization on the size of coresets...
research
07/19/2021

Can we globally optimize cross-validation loss? Quasiconvexity in ridge regression

Models like LASSO and ridge regression are extensively used in practice ...
research
10/06/2019

Ridge Regression: Structure, Cross-Validation, and Sketching

We study the following three fundamental problems about ridge regression...
research
05/07/2020

Fractional ridge regression: a fast, interpretable reparameterization of ridge regression

Ridge regression (RR) is a regularization technique that penalizes the L...
research
07/05/2023

The distribution of Ridgeless least squares interpolators

The Ridgeless minimum ℓ_2-norm interpolator in overparametrized linear r...
research
07/25/2013

Does generalization performance of l^q regularization learning depend on q? A negative example

l^q-regularization has been demonstrated to be an attractive technique i...
research
10/12/2018

Safe Grid Search with Optimal Complexity

Popular machine learning estimators involve regularization parameters th...

Please sign up or login with your details

Forgot password? Click here to reset