A Context-Aware Loss Function for Action Spotting in Soccer Videos

12/03/2019
by   Anthony Cioppa, et al.
16

Action spotting is an important element of general activity understanding. It consists of detecting human-induced events annotated with single timestamps. In this paper, we propose a novel loss function for action spotting. Our loss aims at dealing specifically with the temporal context naturally present around an action. Rather than focusing on the single annotated frame of the action to spot, we consider different temporal segments surrounding it and shape our loss function accordingly. We test our loss on SoccerNet, a large dataset of soccer videos, showing an improvement of 12.8 the generalization capability of our loss function on ActivityNet for activity proposals and detection, by spotting the beginning and the end of each activity. Furthermore, we provide an extended ablation study and identify challenging cases for action spotting in soccer videos. Finally, we qualitatively illustrate how our loss induces a precise temporal understanding of actions, and how such semantic knowledge can be leveraged to design a highlights generator.

READ FULL TEXT

page 1

page 3

page 12

page 16

research
06/20/2019

vireoJD-MM at Activity Detection in Extended Videos

This notebook paper presents an overview and comparative analysis of our...
research
06/07/2019

Detecting the Starting Frame of Actions in Video

To understand causal relationships between events in the world, it is us...
research
09/30/2021

Deep Learning-based Action Detection in Untrimmed Videos: A Survey

Understanding human behavior and activity facilitates advancement of num...
research
04/14/2021

Temporally-Aware Feature Pooling for Action Spotting in Soccer Broadcasts

Toward the goal of automatic production for sports broadcasts, a paramou...
research
03/21/2017

Encouraging LSTMs to Anticipate Actions Very Early

In contrast to the widely studied problem of recognizing an action given...
research
09/30/2017

Unsupervised Segmentation of Action Segments in Egocentric Videos using Gaze

Unsupervised segmentation of action segments in egocentric videos is a d...
research
07/10/2020

Recognition of Instrument-Tissue Interactions in Endoscopic Videos via Action Triplets

Recognition of surgical activity is an essential component to develop co...

Please sign up or login with your details

Forgot password? Click here to reset