In this paper, we conduct a holistic exploration of the Universal
Decomp...
Token dropping is a recently-proposed strategy to speed up the pretraini...
Recent pre-trained language models (PLMs) achieve promising results in
e...
ChatGPT shows remarkable capabilities for machine translation (MT). Seve...
Transfer learning is a simple and powerful method that can be used to bo...
Simultaneous machine translation (SiMT) is usually done via sequence-lev...
Existing research generally treats Chinese character as a minimum unit f...
Pretraining-based (PT-based) automatic evaluation metrics (e.g., BERTSco...
Data augmentations (DA) are the cores to achieving robust
sequence-to-se...
We release 70 small and discriminative test sets for machine translation...
Pre-training (PT) and back-translation (BT) are two simple and powerful
...
The high-quality translation results produced by machine translation (MT...
Previous studies have shown that initializing neural machine translation...
Non-autoregressive translation (NAT) significantly accelerates the infer...
Knowledge distillation (KD) is commonly used to construct synthetic data...
Meta-learning has been sufficiently validated to be beneficial for
low-r...
Encoder layer fusion (EncoderFusion) is a technique to fuse all the enco...
Knowledge distillation (KD) is essential for training non-autoregressive...
Scene text spotting aims to detect and recognize the entire word or sent...
A neural machine translation (NMT) system is expensive to train, especia...
Low-resolution text images are often seen in natural scenes such as docu...
In this paper, we introduce an anchor-box free and single shot instance
...
Word embedding is central to neural machine translation (NMT), which has...
Scene text detection, an essential step of scene text recognition system...
Although deeper and larger neural networks have achieved better performa...
Incidental scene text spotting is considered one of the most difficult a...