Fast FixMatch: Faster Semi-Supervised Learning with Curriculum Batch Size

09/07/2023
by   John Chen, et al.
0

Advances in Semi-Supervised Learning (SSL) have almost entirely closed the gap between SSL and Supervised Learning at a fraction of the number of labels. However, recent performance improvements have often come at the cost of significantly increased training computation. To address this, we propose Curriculum Batch Size (CBS), an unlabeled batch size curriculum which exploits the natural training dynamics of deep neural networks. A small unlabeled batch size is used in the beginning of training and is gradually increased to the end of training. A fixed curriculum is used regardless of dataset, model or number of epochs, and reduced training computations is demonstrated on all settings. We apply CBS, strong labeled augmentation, Curriculum Pseudo Labeling (CPL) <cit.> to FixMatch <cit.> and term the new SSL algorithm Fast FixMatch. We perform an ablation study to show that strong labeled augmentation and/or CPL do not significantly reduce training computations, but, in synergy with CBS, they achieve optimal performance. Fast FixMatch also achieves substantially higher data utilization compared to previous state-of-the-art. Fast FixMatch achieves between 2.1× - 3.4× reduced training computations on CIFAR-10 with all but 40, 250 and 4000 labels removed, compared to vanilla FixMatch, while attaining the same cited state-of-the-art error rate <cit.>. Similar results are achieved for CIFAR-100, SVHN and STL-10. Finally, Fast MixMatch achieves between 2.6× - 3.3× reduced training computations in federated SSL tasks and online/streaming learning SSL tasks, which further demonstrate the generializbility of Fast MixMatch to different scenarios and tasks.

READ FULL TEXT

page 5

page 8

research
10/15/2021

FlexMatch: Boosting Semi-Supervised Learning with Curriculum Pseudo Labeling

The recently proposed FixMatch achieved state-of-the-art results on most...
research
01/16/2020

Curriculum Labeling: Self-paced Pseudo-Labeling for Semi-Supervised Learning

Semi-supervised learning aims to take advantage of a large amount of unl...
research
10/07/2016

Temporal Ensembling for Semi-Supervised Learning

In this paper, we present a simple and efficient method for training dee...
research
06/14/2018

Improving Consistency-Based Semi-Supervised Learning with Weight Averaging

Recent advances in deep unsupervised learning have renewed interest in s...
research
05/17/2023

RelationMatch: Matching In-batch Relationships for Semi-supervised Learning

Semi-supervised learning has achieved notable success by leveraging very...
research
10/28/2022

When does mixup promote local linearity in learned representations?

Mixup is a regularization technique that artificially produces new sampl...
research
01/13/2020

Rethinking Curriculum Learning with Incremental Labels and Adaptive Compensation

Like humans, deep networks learn better when samples are organized and i...

Please sign up or login with your details

Forgot password? Click here to reset