Asymptotic behavior of ℓ_p-based Laplacian regularization in semi-supervised learning

03/02/2016
by   Ahmed El Alaoui, et al.
0

Given a weighted graph with N vertices, consider a real-valued regression problem in a semi-supervised setting, where one observes n labeled vertices, and the task is to label the remaining ones. We present a theoretical study of ℓ_p-based Laplacian regularization under a d-dimensional geometric random graph model. We provide a variational characterization of the performance of this regularized learner as N grows to infinity while n stays constant, the associated optimality conditions lead to a partial differential equation that must be satisfied by the associated function estimate f̂. From this formulation we derive several predictions on the limiting behavior the d-dimensional function f̂, including (a) a phase transition in its smoothness at the threshold p = d + 1, and (b) a tradeoff between smoothness and sensitivity to the underlying unlabeled data distribution P. Thus, over the range p ≤ d, the function estimate f̂ is degenerate and "spiky," whereas for p≥ d+1, the function estimate f̂ is smooth. We show that the effect of the underlying density vanishes monotonically with p, such that in the limit p = ∞, corresponding to the so-called Absolutely Minimal Lipschitz Extension, the estimate f̂ is independent of the distribution P. Under the assumption of semi-supervised smoothness, ignoring P can lead to poor statistical performance, in particular, we construct a specific example for d=1 to demonstrate that p=2 has lower risk than p=∞ due to the former penalty adapting to P and the latter ignoring it. We also provide simulations that verify the accuracy of our predictions for finite sample sizes. Together, these properties show that p = d+1 is an optimal choice, yielding a function estimate f̂ that is both smooth and non-degenerate, while remaining maximally sensitive to P.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/10/2018

Properly-weighted graph Laplacian for semi-supervised learning

The performance of traditional graph Laplacian methods for semi-supervis...
research
07/19/2017

Analysis of p-Laplacian Regularization in Semi-Supervised Learning

We investigate a family of regression problems in a semi-supervised sett...
research
11/28/2011

Adaptive Semisupervised Inference

Semisupervised methods inevitably invoke some assumption that links the ...
research
10/01/2022

Learning Globally Smooth Functions on Manifolds

Smoothness and low dimensional structures play central roles in improvin...
research
04/03/2017

Local nearest neighbour classification with applications to semi-supervised learning

We derive a new asymptotic expansion for the global excess risk of a loc...
research
09/23/2020

Enhancing Mixup-based Semi-Supervised Learning with Explicit Lipschitz Regularization

The success of deep learning relies on the availability of large-scale a...
research
04/07/2012

Density-sensitive semisupervised inference

Semisupervised methods are techniques for using labeled data (X_1,Y_1),....

Please sign up or login with your details

Forgot password? Click here to reset