Introduction to quasi-open set semi-supervised learning for big data analytics

02/04/2020
by   Emile R. Engelbrecht, et al.
0

State-of-the-art performance and low system complexity has made deep-learning an increasingly attractive solution for big data analytics. However, limiting assumptions of end-to-end learning regimes hinder the use of neural networks on large application-grade datasets. This work addresses the assumption that output class-labels are defined for all classes in the domain. The amount of data collected by modern-day sensors span over an incomprehensible range of potential classes. Therefore, we propose a new learning regime where only some, but not all, classes of the training data are of interest to the classification system. The semi-supervised learning scenario in big data requires the assumption of a partial class mismatch between labelled and unlabelled training data. With classification systems required to classify source classes indicated by labelled samples while separating novel classes indicated by unlabelled samples, we find ourselves in an open-set case (vs closed set with only source classes). However, introducing samples from novel classes into the training set indicates a more relaxed open-set case. As such, our proposed regime of quasi-open set semi-supervised learning is introduced. We propose a suitable method to train under quasi-open set semi-supervised learning that makes use of Wasserstein generative adversarial networks (WGANs). A trained classification certainty estimation within the discriminator (or critic) network is used to enable a reject option for the classifier. By placing a threshold on this certainty estimation, the reject option accepts classifications of source classes and rejects novel classes. Big data end-to-end training is promoted by developing models that recognize input samples do not necessarily belong to output labels. We believe this essential for big data analytics, and urge more work under quasi-open set semi-supervised learning.

READ FULL TEXT
research
02/08/2017

An Adversarial Regularisation for Semi-Supervised Training of Structured Output Neural Networks

We propose a method for semi-supervised training of structured-output ne...
research
01/24/2023

Improving Open-Set Semi-Supervised Learning with Self-Supervision

Open-set semi-supervised learning (OSSL) is a realistic setting of semi-...
research
02/06/2021

Open-World Semi-Supervised Learning

Supervised and semi-supervised learning methods have been traditionally ...
research
10/27/2017

A Self-Training Method for Semi-Supervised GANs

Since the creation of Generative Adversarial Networks (GANs), much work ...
research
11/05/2020

Binary classification with ambiguous training data

In supervised learning, we often face with ambiguous (A) samples that ar...
research
05/23/2019

SelectNet: Learning to Sample from the Wild for Imbalanced Data Training

Supervised learning from training data with imbalanced class sizes, a co...
research
06/25/2023

Smart Transformation of EFL Teaching and Learning Approaches

The calibration of the EFL teaching and learning approaches with Artific...

Please sign up or login with your details

Forgot password? Click here to reset