Scalable Semi-Supervised SVM via Triply Stochastic Gradients

07/26/2019
by   Xiang Geng, et al.
2

Semi-supervised learning (SSL) plays an increasingly important role in the big data era because a large number of unlabeled samples can be used effectively to improve the performance of the classifier. Semi-supervised support vector machine (S^3VM) is one of the most appealing methods for SSL, but scaling up S^3VM for kernel learning is still an open problem. Recently, a doubly stochastic gradient (DSG) algorithm has been proposed to achieve efficient and scalable training for kernel methods. However, the algorithm and theoretical analysis of DSG are developed based on the convexity assumption which makes them incompetent for non-convex problems such as S^3VM. To address this problem, in this paper, we propose a triply stochastic gradient algorithm for S^3VM, called TSGS^3VM. Specifically, to handle two types of data instances involved in S^3VM, TSGS^3VM samples a labeled instance and an unlabeled instance as well with the random features in each iteration to compute a triply stochastic gradient. We use the approximated gradient to update the solution. More importantly, we establish new theoretic analysis for TSGS^3VM which guarantees that TSGS^3VM can converge to a stationary point. Extensive experimental results on a variety of datasets demonstrate that TSGS^3VM is much more efficient and scalable than existing S^3VM algorithms.

READ FULL TEXT
research
07/29/2019

Quadruply Stochastic Gradients for Large Scale Nonlinear Semi-Supervised AUC Optimization

Semi-supervised learning is pervasive in real-world applications, where ...
research
12/24/2019

Quadruply Stochastic Gradient Method for Large Scale Nonlinear Semi-Supervised Ordinal Regression AUC Optimization

Semi-supervised ordinal regression (S^2OR) problems are ubiquitous in re...
research
04/22/2022

Quantum Semi-Supervised Kernel Learning

Quantum computing leverages quantum effects to build algorithms that are...
research
04/21/2018

Stability of the Stochastic Gradient Method for an Approximated Large Scale Kernel Machine

In this paper we measured the stability of stochastic gradient method (S...
research
07/07/2022

FastHebb: Scaling Hebbian Training of Deep Neural Networks to ImageNet Level

Learning algorithms for Deep Neural Networks are typically based on supe...
research
06/07/2020

Optimally Combining Classifiers for Semi-Supervised Learning

This paper considers semi-supervised learning for tabular data. It is wi...
research
07/22/2018

PaloBoost: An Overfitting-robust TreeBoost with Out-of-Bag Sample Regularization Techniques

Stochastic Gradient TreeBoost is often found in many winning solutions i...

Please sign up or login with your details

Forgot password? Click here to reset