You were saying? – Spoken Language in the V3C Dataset

12/15/2022
by   Luca Rossetto, et al.
0

This paper presents an analysis of the distribution of spoken language in the V3C video retrieval benchmark dataset based on automatically generated transcripts. It finds that a large portion of the dataset is covered by spoken language. Since language transcripts can be quickly and accurately described, this has implications for retrieval tasks such as known-item search.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/10/2021

VT-SSum: A Benchmark Dataset for Video Transcript Segmentation and Summarization

Video transcript summarization is a fundamental task for video understan...
research
09/23/2020

The importance of fillers for text representations of speech transcripts

While being an essential component of spoken language, fillers (e.g."um"...
research
10/14/2021

Spoken ObjectNet: A Bias-Controlled Spoken Caption Dataset

Visually-grounded spoken language datasets can enable models to learn cr...
research
11/02/2020

IndoLEM and IndoBERT: A Benchmark Dataset and Pre-trained Language Model for Indonesian NLP

Although the Indonesian language is spoken by almost 200 million people ...
research
09/19/2023

Multimodal Modeling For Spoken Language Identification

Spoken language identification refers to the task of automatically predi...
research
06/20/2023

MSVD-Indonesian: A Benchmark for Multimodal Video-Text Tasks in Indonesian

Multimodal learning on video and text data has been receiving growing at...
research
09/23/2022

Cem Mil Podcasts: A Spoken Portuguese Document Corpus

This document describes the Portuguese language podcast dataset released...

Please sign up or login with your details

Forgot password? Click here to reset