Optimal Sketching Bounds for Sparse Linear Regression

04/05/2023
by   Tung Mai, et al.
0

We study oblivious sketching for k-sparse linear regression under various loss functions such as an ℓ_p norm, or from a broad class of hinge-like loss functions, which includes the logistic and ReLU losses. We show that for sparse ℓ_2 norm regression, there is a distribution over oblivious sketches with Θ(klog(d/k)/ε^2) rows, which is tight up to a constant factor. This extends to ℓ_p loss with an additional additive O(klog(k/ε)/ε^2) term in the upper bound. This establishes a surprising separation from the related sparse recovery problem, which is an important special case of sparse regression. For this problem, under the ℓ_2 norm, we observe an upper bound of O(k log (d)/ε + klog(k/ε)/ε^2) rows, showing that sparse recovery is strictly easier to sketch than sparse regression. For sparse regression under hinge-like loss functions including sparse logistic and sparse ReLU regression, we give the first known sketching bounds that achieve o(d) rows showing that O(μ^2 klog(μ n d/ε)/ε^2) rows suffice, where μ is a natural complexity parameter needed to obtain relative error bounds for these loss functions. We again show that this dimension is tight, up to lower order terms and the dependence on μ. Finally, we show that similar sketching bounds can be achieved for LASSO regression, a popular convex relaxation of sparse regression, where one aims to minimize Ax-b_2^2+λx_1 over x∈ℝ^d. We show that sketching dimension O(log(d)/(λε)^2) suffices and that the dependence on d and λ is tight.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/09/2021

Active Sampling for Linear Regression Beyond the ℓ_2 Norm

We study active sampling algorithms for linear regression, which aim to ...
research
10/26/2020

Estimation of the l_2-norm and testing in sparse linear regression with unknown variance

We consider the related problems of estimating the l_2-norm and the squa...
research
10/07/2018

Error bounds for sparse classifiers in high-dimensions

We prove an L2 recovery bound for a family of sparse estimators defined ...
research
12/03/2020

Sample-efficient L0-L2 constrained structure learning of sparse Ising models

We consider the problem of learning the underlying graph of a sparse Isi...
research
02/11/2022

Scale-free Unconstrained Online Learning for Curved Losses

A sequence of works in unconstrained online convex optimisation have inv...
research
02/17/2020

Sharp Asymptotics and Optimal Performance for Inference in Binary Models

We study convex empirical risk minimization for high-dimensional inferen...
research
02/22/2019

Nonconvex sampling with the Metropolis-adjusted Langevin algorithm

The Langevin Markov chain algorithms are widely deployed methods to samp...

Please sign up or login with your details

Forgot password? Click here to reset