A common training technique for language models is teacher forcing (TF)....
Conformer, a convolution-augmented Transformer variant, has become the d...
Self-supervised speech representation learning (SSL) has shown to be
eff...
Spoken language understanding (SLU) tasks have been studied for many dec...
Self-supervised pre-trained transformers have improved the state of the ...
Conformer, combining convolution and self-attention sequentially to capt...
We introduce Wav2Seq, the first self-supervised approach to pre-train bo...
Spoken language understanding (SLU) tasks involve mapping from speech au...
Progress in speech processing has been facilitated by shared datasets an...
This paper is a study of performance-efficiency trade-offs in pre-traine...
Automatic speech recognition (ASR) models make fewer errors when more
su...
Most computer science conferences rely on paper bidding to assign review...
With rapid progress across platforms for quantum systems, the problem of...
We study the problem of few-sample fine-tuning of BERT contextual
repres...
Modern neural network training relies heavily on data augmentation for
i...
Although according to several benchmarks automatic machine reading
compr...
A widely deployed method for reducing the training time of deep neural
n...
We propose BERTScore, an automatic evaluation metric for text generation...
In this technical report, we introduce FastFusionNet, an efficient varia...
Graph Convolutional Networks (GCNs) and their variants have experienced
...
Self-attention is a useful mechanism to build generative models for lang...
Graph convolutional networks (GCNs) have been widely used for classifyin...
Evaluating generative adversarial networks (GANs) is inherently challeng...
State-of-the-art deep reading comprehension models are dominated by recu...
The machine learning community has become increasingly concerned with th...