The USYD-JD Speech Translation System for IWSLT 2021

07/24/2021
by   Liang Ding, et al.
0

This paper describes the University of Sydney JD's joint submission of the IWSLT 2021 low resource speech translation task. We participated in the Swahili-English direction and got the best scareBLEU (25.3) score among all the participants. Our constrained system is based on a pipeline framework, i.e. ASR and NMT. We trained our models with the officially provided ASR and MT datasets. The ASR system is based on the open-sourced tool Kaldi and this work mainly explores how to make the most of the NMT models. To reduce the punctuation errors generated by the ASR model, we employ our previous work SlotRefine to train a punctuation correction model. To achieve better translation performance, we explored the most recent effective strategies, including back translation, knowledge distillation, multi-feature reranking and transductive finetuning. For model structure, we tried auto-regressive and non-autoregressive models, respectively. In addition, we proposed two novel pre-train approaches, i.e. de-noising training and bidirectional training to fully exploit the data. Extensive experiments show that adding the above techniques consistently improves the BLEU scores, and the final submission system outperforms the baseline (Transformer ensemble model trained with the original parallel data) by approximately 10.8 BLEU score, achieving the SOTA performance.

READ FULL TEXT
research
06/30/2019

The University of Sydney's Machine Translation System for WMT19

This paper describes the University of Sydney's submission of the WMT 20...
research
06/30/2021

IMS' Systems for the IWSLT 2021 Low-Resource Speech Translation Task

This paper describes the submission to the IWSLT 2021 Low-Resource Speec...
research
04/13/2021

Source and Target Bidirectional Knowledge Distillation for End-to-end Speech Translation

A conventional approach to improving the performance of end-to-end speec...
research
08/09/2021

The HW-TSC's Offline Speech Translation Systems for IWSLT 2021 Evaluation

This paper describes our work in participation of the IWSLT-2021 offline...
research
11/29/2022

CUNI Submission in WMT22 General Task

We present the CUNI-Bergamot submission for the WMT22 General translatio...
research
11/02/2018

Improving the Robustness of Speech Translation

Although neural machine translation (NMT) has achieved impressive progre...
research
10/18/2022

Simultaneous Translation for Unsegmented Input: A Sliding Window Approach

In the cascaded approach to spoken language translation (SLT), the ASR o...

Please sign up or login with your details

Forgot password? Click here to reset