On Discarding, Caching, and Recalling Samples in Active Learning

by   Ashish Kapoor, et al.

We address challenges of active learning under scarce informational resources in non-stationary environments. In real-world settings, data labeled and integrated into a predictive model may become invalid over time. However, the data can become informative again with switches in context and such changes may indicate unmodeled cyclic or other temporal dynamics. We explore principles for discarding, caching, and recalling labeled data points in active learning based on computations of value of information. We review key concepts and study the value of the methods via investigations of predictive performance and costs of acquiring data for simulated and real-world data sets.


page 1

page 2

page 3

page 4


One-Round Active Learning

Active learning has been a main solution for reducing data labeling cost...

Low-Regret Active learning

We develop an online learning algorithm for identifying unlabeled data p...

Graph-based Reinforcement Learning for Active Learning in Real Time: An Application in Modeling River Networks

Effective training of advanced ML models requires large amounts of label...

Combining self-labeling and demand based active learning for non-stationary data streams

Learning from non-stationary data streams is a research direction that g...

Active Learning-based Classification in Automated Connected Vehicles

Machine learning has emerged as a promising paradigm for enabling connec...

Active Learning for Regression with Aggregated Outputs

Due to the privacy protection or the difficulty of data collection, we c...

Prediction-Oriented Bayesian Active Learning

Information-theoretic approaches to active learning have traditionally f...

Please sign up or login with your details

Forgot password? Click here to reset