Improving Cascaded Unsupervised Speech Translation with Denoising Back-translation

05/12/2023
by   Yu-Kuan Fu, et al.
0

Most of the speech translation models heavily rely on parallel data, which is hard to collect especially for low-resource languages. To tackle this issue, we propose to build a cascaded speech translation system without leveraging any kind of paired data. We use fully unpaired data to train our unsupervised systems and evaluate our results on CoVoST 2 and CVSS. The results show that our work is comparable with some other early supervised methods in some language pairs. While cascaded systems always suffer from severe error propagation problems, we proposed denoising back-translation (DBT), a novel approach to building robust unsupervised neural machine translation (UNMT). DBT successfully increases the BLEU score by 0.7–0.9 in all three translation directions. Moreover, we simplified the pipeline of our cascaded system to reduce inference latency and conducted a comprehensive analysis of every part of our work. We also demonstrate our unsupervised speech translation results on the established website.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/18/2022

Simple and Effective Unsupervised Speech Translation

The amount of labeled data to train models for speech tasks is limited f...
research
04/07/2020

Unsupervised Neural Machine Translation with Indirect Supervision

Neural machine translation (NMT) is ineffective for zero-resource langua...
research
10/24/2022

Does Joint Training Really Help Cascaded Speech Translation?

Currently, in speech translation, the straightforward approach - cascadi...
research
05/27/2023

Translatotron 3: Speech to Speech Translation with Monolingual Data

This paper presents Translatotron 3, a novel approach to train a direct ...
research
10/20/2021

Multilingual Unsupervised Neural Machine Translation with Denoising Adapters

We consider the problem of multilingual unsupervised machine translation...
research
06/11/2019

Translating Translationese: A Two-Step Approach to Unsupervised Machine Translation

Given a rough, word-by-word gloss of a source language sentence, target ...
research
09/27/2021

Integrated Training for Sequence-to-Sequence Models Using Non-Autoregressive Transformer

Complex natural language applications such as speech translation or pivo...

Please sign up or login with your details

Forgot password? Click here to reset