ByT5 model for massively multilingual grapheme-to-phoneme conversion

04/06/2022
by   Jian Zhu, et al.
0

In this study, we tackle massively multilingual grapheme-to-phoneme conversion through implementing G2P models based on ByT5. We have curated a G2P dataset from various sources that covers around 100 languages and trained large-scale multilingual G2P models based on ByT5. We found that ByT5 operating on byte-level inputs significantly outperformed the token-based mT5 model in terms of multilingual G2P. Pairwise comparison with monolingual models in these languages suggests that multilingual ByT5 models generally lower the phone error rate by jointly learning from a variety of languages. The pretrained model can further benefit low resource G2P through zero-shot prediction on unseen languages or provides pretrained weights for finetuning, which helps the model converge to a lower phone error rate than randomly initialized weights. To facilitate future research on multilingual G2P, we make available our code and pretrained multilingual G2P models at: https://github.com/lingjzhu/CharsiuG2P.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/14/2020

Deep Learning Models for Multilingual Hate Speech Detection

Hate speech detection is a challenging problem with most of the datasets...
research
04/03/2019

75 Languages, 1 Model: Parsing Universal Dependencies Universally

We present UDify, a multilingual multi-task model capable of accurately ...
research
02/26/2020

Universal Phone Recognition with a Multilingual Allophone System

Multilingual models can improve language processing, particularly for lo...
research
09/15/2023

How Transferable are Attribute Controllers on Pretrained Multilingual Translation Models?

Customizing machine translation models to comply with fine-grained attri...
research
08/04/2017

Massively Multilingual Neural Grapheme-to-Phoneme Conversion

Grapheme-to-phoneme conversion (g2p) is necessary for text-to-speech and...
research
06/23/2020

One Model to Pronounce Them All: Multilingual Grapheme-to-Phoneme Conversion With a Transformer Ensemble

The task of grapheme-to-phoneme (G2P) conversion is important for both s...
research
05/23/2023

FOCUS: Effective Embedding Initialization for Specializing Pretrained Multilingual Models on a Single Language

Using model weights pretrained on a high-resource language as a warm sta...

Please sign up or login with your details

Forgot password? Click here to reset