We introduce HK-LegiCoST, a new three-way parallel corpus of
Cantonese-E...
Large language models such as BERT and the GPT series started a paradigm...
Multilingual machine translation has proven immensely useful for low-res...
The phenomena of in-context learning has typically been thought of as
"l...
Bilingual lexicons form a critical component of various natural language...
The ability to extract high-quality translation dictionaries from monoli...
The advent of transformer-based models such as BERT has led to the rise ...
Much recent work in bilingual lexicon induction (BLI) views word embeddi...
This article describes an efficient end-to-end speech translation (E2E-S...
This paper describes the ESPnet-ST group's IWSLT 2021 submission in the
...
In the field of machine learning, the well-trained model is assumed to b...
"Transcription bottlenecks", created by a shortage of effective human
tr...
Fast inference speed is an important goal towards real-world deployment ...
We explore the application of very deep Transformer models for Neural Ma...
Learning to rank is an important task that has been successfully deploye...
We present ESPnet-ST, which is designed for the quick development of
spe...
Despite the reported success of unsupervised machine translation (MT), t...
We explore best practices for training small, memory efficient machine
t...
Adapting machine translation systems in the real world is a difficult
pr...
Universal feature extractors, such as BERT for natural language processi...
Sequence-level knowledge distillation (SLKD) is a model compression tech...
In this paper, we propose a simple yet effective framework for multiling...
We unify different broad-coverage semantic parsing tasks under a transdu...
Most neural machine translation systems are built upon subword units
ext...
We propose an attention-based model that treats AMR parsing as
sequence-...
We introduce a curriculum learning approach to adapt generic neural mach...
Community question-answering (CQA) platforms have become very popular fo...
Data privacy is an important issue for "machine learning as a service"
p...
Machine translation systems based on deep neural networks are expensive ...
We present a large-scale dataset, ReCoRD, for machine reading comprehens...
This paper presents an extension of the Stochastic Answer Network (SAN),...
To better understand the effectiveness of continued training, we analyze...
Standard neural machine translation (NMT) systems operate primarily on w...
Neural Machine Translation (NMT) in low-resource settings and of
morphol...
Using pre-trained word embeddings as input layer is a common practice in...
Cross-lingual information extraction (CLIE) is an important and challeng...
We introduce the task of cross-lingual semantic parsing: mapping content...
Fine-grained entity typing is the task of assigning fine-grained semanti...
We propose a stochastic answer network (SAN) to explore multi-step infer...
We propose a simple yet robust stochastic answer network (SAN) that simu...
Reading comprehension (RC) is a challenging task that requires synthesis...
We develop a streaming (one-pass, bounded-memory) word embedding algorit...
We describe DyNet, a toolkit for implementing neural network models base...
Humans have the capacity to draw common-sense inferences from natural
la...
Language processing mechanism by humans is generally more robust than
co...
We propose a transition-based dependency parser using Recurrent Neural
N...
In this short note, we present an extension of long short-term memory (L...
We investigate the hypothesis that word representations ought to incorpo...