Socratic Learning: Augmenting Generative Models to Incorporate Latent Subsets in Training Data

10/25/2016
∙
by   Paroma Varma, et al.
∙
0
∙

A challenge in training discriminative models like neural networks is obtaining enough labeled training data. Recent approaches use generative models to combine weak supervision sources, like user-defined heuristics or knowledge bases, to label training data. Prior work has explored learning accuracies for these sources even without ground truth labels, but they assume that a single accuracy parameter is sufficient to model the behavior of these sources over the entire training set. In particular, they fail to model latent subsets in the training data in which the supervision sources perform differently than on average. We present Socratic learning, a paradigm that uses feedback from a corresponding discriminative model to automatically identify these subsets and augments the structure of the generative model accordingly. Experimentally, we show that without any ground truth labels, the augmented generative model reduces error by up to 56.06 state-of-the-art weak supervision technique that utilizes generative models.

READ FULL TEXT
research
∙ 09/07/2017

Inferring Generative Model Structure with Static Analysis

Obtaining enough labeled data to robustly train complex discriminative m...
research
∙ 05/25/2016

Data Programming: Creating Large Training Sets, Quickly

Large labeled training sets are the critical building blocks of supervis...
research
∙ 07/27/2022

Learning Hyper Label Model for Programmatic Weak Supervision

To reduce the human annotation efforts, the programmatic weak supervisio...
research
∙ 03/02/2017

Learning the Structure of Generative Models without Labeled Data

Curating labeled training data has become the primary bottleneck in mach...
research
∙ 07/05/2021

End-to-End Weak Supervision

Aggregating multiple sources of weak supervision (WS) can ease the data-...
research
∙ 06/08/2021

Learning from Multiple Noisy Partial Labelers

Programmatic weak supervision creates models without hand-labeled traini...
research
∙ 02/27/2020

Fast and Three-rious: Speeding Up Weak Supervision with Triplet Methods

Weak supervision is a popular method for building machine learning model...

Please sign up or login with your details

Forgot password? Click here to reset