Large Scale Graph Learning from Smooth Signals

by   Vassilis Kalofolias, et al.

Graphs are a prevalent tool in data science, as they model the inherent structure of the data. They have been used successfully in unsupervised and semi-supervised learning. Typically they are constructed either by connecting nearest samples, or by learning them from data, solving an optimization problem. While graph learning does achieve a better quality, it also comes with a higher computational cost. In particular, the current state-of-the-art model cost is O(n^2) for n samples. In this paper, we show how to scale it, obtaining an approximation with leading cost of O(n(n)), with quality that approaches the exact graph learning model. Our algorithm uses known approximate nearest neighbor techniques to reduce the number of variables, and automatically selects the correct parameters of the model, requiring a single intuitive input: the desired edge density.


page 1

page 2

page 3

page 4


Semi-supervised Learning with Density Based Distances

We present a simple, yet effective, approach to Semi-Supervised Learning...

Graph based Nearest Neighbor Search: Promises and Failures

Recently, graph based nearest neighbor search gets more and more popular...

A soft nearest-neighbor framework for continual semi-supervised learning

Despite significant advances, the performance of state-of-the-art contin...

Solving Traveling Salesman Problem by Marker Method

In this paper we use marker method and propose a new mutation operator t...

Incremental Spectral Sparsification for Large-Scale Graph-Based Semi-Supervised Learning

While the harmonic function solution performs well in many semi-supervis...

Beyond kNN: Adaptive, Sparse Neighborhood Graphs via Optimal Transport

Nearest neighbour graphs are widely used to capture the geometry or topo...

Please sign up or login with your details

Forgot password? Click here to reset