DeepAI AI Chat
Log In Sign Up

Exploiting Context Information for Generic Event Boundary Captioning

by   Jinrui Zhang, et al.
Southern University of Science & Technology
The University of Hong Kong

Generic Event Boundary Captioning (GEBC) aims to generate three sentences describing the status change for a given time boundary. Previous methods only process the information of a single boundary at a time, which lacks utilization of video context information. To tackle this issue, we design a model that directly takes the whole video as input and generates captions for all boundaries parallelly. The model could learn the context information for each time boundary by modeling the boundary-boundary interactions. Experiments demonstrate the effectiveness of context information. The proposed method achieved a 72.84 score on the test set, and we reached the 2^nd place in this challenge. Our code is available at: <>


Dual-Stream Transformer for Generic Event Boundary Captioning

This paper describes our champion solution for the CVPR2022 Generic Even...

REVECA – Rich Encoder-decoder framework for Video Event CAptioner

We describe an approach used in the Generic Boundary Event Captioning ch...

Generic Event Boundary Captioning: A Benchmark for Status Changes Understanding

Cognitive science has shown that humans perceive videos in terms of even...

Winning the CVPR'2021 Kinetics-GEBD Challenge: Contrastive Learning Approach

Generic Event Boundary Detection (GEBD) is a newly introduced task that ...

Masked Autoencoders for Generic Event Boundary Detection CVPR'2022 Kinetics-GEBD Challenge

Generic Event Boundary Detection (GEBD) tasks aim at detecting generic, ...