BSL-1K: Scaling up co-articulated sign language recognition using mouthing cues

07/23/2020
by   Samuel Albanie, et al.
2

Recent progress in fine-grained gesture and action classification, and machine translation, point to the possibility of automated sign language recognition becoming a reality. A key stumbling block in making progress towards this goal is a lack of appropriate training data, stemming from the high complexity of sign annotation and a limited supply of qualified annotators. In this work, we introduce a new scalable approach to data collection for sign recognition in continuous videos. We make use of weakly-aligned subtitles for broadcast footage together with a keyword spotting method to automatically localise sign-instances for a vocabulary of 1,000 signs in 1,000 hours of video. We make the following contributions: (1) We show how to use mouthing cues from signers to obtain high-quality annotations from video data - the result is the BSL-1K dataset, a collection of British Sign Language (BSL) signs of unprecedented scale; (2) We show that we can use BSL-1K to train strong sign recognition models for co-articulated signs in BSL and that these models additionally form excellent pretraining for other sign languages and benchmarks - we exceed the state of the art on both the MSASL and WLASL benchmarks. Finally, (3) we propose new large-scale evaluation sets for the tasks of sign recognition and sign spotting and provide baselines which we hope will serve to stimulate research in this area.

READ FULL TEXT

page 5

page 7

page 13

page 20

page 21

page 23

page 26

research
03/30/2021

Read and Attend: Temporal Localisation in Sign Language Videos

The objective of this work is to annotate sign instances across a broad ...
research
08/04/2022

Automatic dense annotation of large-vocabulary sign language videos

Recently, sign language researchers have turned to sign language interpr...
research
09/01/2022

Topic Detection in Continuous Sign Language Videos

Significant progress has been made recently on challenging tasks in auto...
research
05/06/2021

Aligning Subtitles in Sign Language Videos

The goal of this work is to temporally align asynchronous subtitles in s...
research
11/16/2022

Weakly-supervised Fingerspelling Recognition in British Sign Language Videos

The goal of this work is to detect and recognize sequences of letters si...
research
04/10/2020

ASL Recognition with Metric-Learning based Lightweight Network

In the past decades the set of human tasks that are solved by machines w...
research
10/12/2021

OpenHands: Making Sign Language Recognition Accessible with Pose-based Pretrained Models across Languages

AI technologies for Natural Languages have made tremendous progress rece...

Please sign up or login with your details

Forgot password? Click here to reset