Keyframe Segmentation and Positional Encoding for Video-guided Machine Translation Challenge 2020

06/23/2020
by   Tosho Hirasawa, et al.
0

Video-guided machine translation as one of multimodal neural machine translation tasks targeting on generating high-quality text translation by tangibly engaging both video and text. In this work, we presented our video-guided machine translation system in approaching the Video-guided Machine Translation Challenge 2020. This system employs keyframe-based video feature extractions along with the video feature positional encoding. In the evaluation phase, our system scored 36.60 corpus-level BLEU-4 and achieved the 1st place on the Video-guided Machine Translation Challenge 2020.

READ FULL TEXT
research
04/08/2021

Extended Parallel Corpus for Amharic-English Machine Translation

This paper describes the acquisition, preprocessing, segmentation, and a...
research
11/28/2019

Multimodal Machine Translation through Visuals and Speech

Multimodal machine translation involves drawing information from more th...
research
01/09/2023

Applying Automated Machine Translation to Educational Video Courses

We studied the capability of automated machine translation in the online...
research
10/21/2021

Video and Text Matching with Conditioned Embeddings

We present a method for matching a text sentence from a given corpus to ...
research
09/19/2022

A Snapshot into the Possibility of Video Game Machine Translation

We present in this article what we believe to be one of the first attemp...
research
08/20/2019

A Lost Croatian Cybernetic Machine Translation Program

We are exploring the historical significance of research in the field of...
research
12/22/2014

Bayesian Optimisation for Machine Translation

This paper presents novel Bayesian optimisation algorithms for minimum e...

Please sign up or login with your details

Forgot password? Click here to reset