Generalization Error without Independence: Denoising, Linear Regression, and Transfer Learning

05/26/2023
by   Chinmaya Kausik, et al.
0

Studying the generalization abilities of linear models with real data is a central question in statistical learning. While there exist a limited number of prior important works (Loureiro et al. (2021A, 2021B), Wei et al. 2022) that do validate theoretical work with real data, these works have limitations due to technical assumptions. These assumptions include having a well-conditioned covariance matrix and having independent and identically distributed data. These assumptions are not necessarily valid for real data. Additionally, prior works that do address distributional shifts usually make technical assumptions on the joint distribution of the train and test data (Tripuraneni et al. 2021, Wu and Xu 2020), and do not test on real data. In an attempt to address these issues and better model real data, we look at data that is not I.I.D. but has a low-rank structure. Further, we address distributional shift by decoupling assumptions on the training and test distribution. We provide analytical formulas for the generalization error of the denoising problem that are asymptotically exact. These are used to derive theoretical results for linear regression, data augmentation, principal component regression, and transfer learning. We validate all of our theoretical results on real data and have a low relative mean squared error of around 1 between the empirical risk and our estimated risk.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/20/2021

Distributional Anchor Regression

Prediction models often fail if train and test data do not stem from the...
research
06/04/2021

Out-of-Distribution Generalization in Kernel Regression

In real word applications, data generating process for training a machin...
research
02/18/2021

Transfer Learning for Linear Regression: a Statistical Test of Gain

Transfer learning, also referred as knowledge transfer, aims at reusing ...
research
07/17/2023

Revisiting the Robustness of the Minimum Error Entropy Criterion: A Transfer Learning Case Study

Coping with distributional shifts is an important part of transfer learn...
research
09/12/2019

A comparison of some conformal quantile regression methods

We compare two recently proposed methods that combine ideas from conform...
research
10/03/2022

Understanding Influence Functions and Datamodels via Harmonic Analysis

Influence functions estimate effect of individual data points on predict...
research
04/30/2020

Generalization Error for Linear Regression under Distributed Learning

Distributed learning facilitates the scaling-up of data processing by di...

Please sign up or login with your details

Forgot password? Click here to reset