DeepAI AI Chat
Log In Sign Up

A Simple Baseline for Batch Active Learning with Stochastic Acquisition Functions

by   Andreas Kirsch, et al.

In active learning, new labels are commonly acquired in batches. However, common acquisition functions are only meant for one-sample acquisition rounds at a time, and when their scores are used naively for batch acquisition, they result in batches lacking diversity, which deteriorates performance. On the other hand, state-of-the-art batch acquisition functions are costly to compute. In this paper, we present a novel class of stochastic acquisition functions that extend one-sample acquisition functions to the batch setting by observing how one-sample acquisition scores change as additional samples are acquired and modelling this difference for additional batch samples. We simply acquire new samples by sampling from the pool set using a Gibbs distribution based on the acquisition scores. Our acquisition functions are both vastly cheaper to compute and out-perform other batch acquisition functions.


page 1

page 2

page 3

page 4


Diversity Enhanced Active Learning with Strictly Proper Scoring Rules

We study acquisition functions for active learning (AL) for text classif...

Active Multi-Information Source Bayesian Quadrature

Bayesian quadrature (BQ) is a sample-efficient probabilistic numerical m...

BatchBALD: Efficient and Diverse Batch Acquisition for Deep Bayesian Active Learning

We develop BatchBALD, a tractable approximation to the mutual informatio...

Actively Learning Deep Neural Networks with Uncertainty Sampling Based on Sum-Product Networks

Active learning is popular approach for reducing the amount of data in t...

Deep active learning for nonlinear system identification

The exploding research interest for neural networks in modeling nonlinea...

Truly Batch Apprenticeship Learning with Deep Successor Features

We introduce a novel apprenticeship learning algorithm to learn an exper...