Much Ado About Time: Exhaustive Annotation of Temporal Data

07/25/2016
by   Gunnar A. Sigurdsson, et al.
0

Large-scale annotated datasets allow AI systems to learn from and build upon the knowledge of the crowd. Many crowdsourcing techniques have been developed for collecting image annotations. These techniques often implicitly rely on the fact that a new input image takes a negligible amount of time to perceive. In contrast, we investigate and determine the most cost-effective way of obtaining high-quality multi-label annotations for temporal data such as videos. Watching even a short 30-second video clip requires a significant time investment from a crowd worker; thus, requesting multiple annotations following a single viewing is an important cost-saving strategy. But how many questions should we ask per video? We conclude that the optimal strategy is to ask as many questions as possible in a HIT (up to 52 binary questions after watching a 30-second video clip in our experiments). We demonstrate that while workers may not correctly answer all questions, the cost-benefit analysis nevertheless favors consensus from multiple such cheap-yet-imperfect iterations over more complex alternatives. When compared with a one-question-per-video baseline, our method is able to achieve a 10 baseline) at comparable precision (83.8 half the annotation time (3.8 minutes ours compared to 7.1 minutes baseline). We demonstrate the effectiveness of our method by collecting multi-label annotations of 157 human activities on 1,815 videos.

READ FULL TEXT
research
11/07/2019

Active Multi-Label Crowd Consensus

Crowdsourcing is an economic and efficient strategy aimed at collecting ...
research
09/06/2021

Rethinking Crowdsourcing Annotation: Partial Annotation with Salient Labels for Multi-Label Image Classification

Annotated images are required for both supervised model training and eva...
research
05/13/2019

VideoGraph: Recognizing Minutes-Long Human Activities in Videos

Many human activities take minutes to unfold. To represent them, related...
research
12/01/2020

Just Ask: Learning to Answer Questions from Millions of Narrated Videos

Modern approaches to visual question answering require large annotated d...
research
09/17/2018

Crowd-Assisted Polyp Annotation of Virtual Colonoscopy Videos

Virtual colonoscopy (VC) allows a radiologist to navigate through a 3D c...
research
02/06/2023

Interface Design for Crowdsourcing Hierarchical Multi-Label Text Annotations

Human data labeling is an important and expensive task at the heart of s...
research
09/29/2020

CrowdMOT: Crowdsourcing Strategies for Tracking Multiple Objects in Videos

Crowdsourcing is a valuable approach for tracking objects in videos in a...

Please sign up or login with your details

Forgot password? Click here to reset