DeepAI AI Chat
Log In Sign Up

Feature space approximation for kernel-based supervised learning

by   Patrick Gelß, et al.
Freie Universität Berlin

We propose a method for the approximation of high- or even infinite-dimensional feature vectors, which play an important role in supervised learning. The goal is to reduce the size of the training data, resulting in lower storage consumption and computational complexity. Furthermore, the method can be regarded as a regularization technique, which improves the generalizability of learned target functions. We demonstrate significant improvements in comparison to the computation of data-driven predictions involving the full training data set. The method is applied to classification and regression problems from different application areas such as image recognition, system identification, and oceanographic time series analysis.


Improvements to Supervised EM Learning of Shared Kernel Models by Feature Space Partitioning

Expectation maximisation (EM) is usually thought of as an unsupervised l...

Kernel-based Prediction of Non-Markovian Time Series

A nonparametric method to predict non-Markovian time series of partially...

Identifying Mislabeled Images in Supervised Learning Utilizing Autoencoder

Supervised learning is based on the assumption that the ground truth in ...

Estimating the time-lapse between medical insurance reimbursement with non-parametric regression models

Non-parametric supervised learning algorithms represent a succinct class...

Improved graph-based SFA: Information preservation complements the slowness principle

Slow feature analysis (SFA) is an unsupervised-learning algorithm that e...

Coloring black boxes: visualization of neural network decisions

Neural networks are commonly regarded as black boxes performing incompre...

Kernel-Based Ensemble Learning in Python

We propose a new supervised learning algorithm, for classification and r...