Extreme clicking for efficient object annotation

08/09/2017
by   Dim P. Papadopoulos, et al.
0

Manually annotating object bounding boxes is central to building computer vision datasets, and it is very time consuming (annotating ILSVRC [53] took 35s for one high-quality box [62]). It involves clicking on imaginary corners of a tight box around the object. This is difficult as these corners are often outside the actual object and several adjustments are required to obtain a tight box. We propose extreme clicking instead: we ask the annotator to click on four physical points on the object: the top, bottom, left- and right-most points. This task is more natural and these points are easy to find. We crowd-source extreme point annotations for PASCAL VOC 2007 and 2012 and show that (1) annotation time is only 7s per box, 5x faster than the traditional way of drawing boxes [62]; (2) the quality of the boxes is as good as the original ground-truth drawn the traditional way; (3) detectors trained on our annotations are as accurate as those trained on the original ground-truth. Moreover, our extreme clicking strategy not only yields box coordinates, but also four accurate boundary points. We show (4) how to incorporate them into GrabCut to obtain more accurate segmentations than those delivered when initializing it from bounding boxes; (5) semantic segmentations models trained on these segmentations outperform those trained on segmentations derived from bounding boxes.

READ FULL TEXT

page 1

page 3

page 4

page 5

page 6

page 10

research
02/26/2016

We don't need no bounding-boxes: Training object class detectors using only human verification

Training object class detectors typically requires a large set of images...
research
07/16/2018

Leveraging Pre-Trained 3D Object Detection Models For Fast Ground Truth Generation

Training 3D object detectors for autonomous driving has been limited to ...
research
01/28/2022

Self-paced learning to improve text row detection in historical documents with missing labels

An important preliminary step of optical character recognition systems i...
research
03/26/2023

CRRS: Concentric Rectangles Regression Strategy for Multi-point Representation on Fisheye Images

Modern object detectors take advantage of rectangular bounding boxes as ...
research
05/05/2021

A Step Toward More Inclusive People Annotations for Fairness

The Open Images Dataset contains approximately 9 million images and is a...
research
09/21/2019

Generating Positive Bounding Boxes for Balanced Training of Object Detectors

Two-stage deep object detectors generate a set of regions-of-interest (R...
research
05/25/2019

Efficient Object Annotation via Speaking and Pointing

Deep neural networks deliver state-of-the-art visual recognition, but th...

Please sign up or login with your details

Forgot password? Click here to reset