Regularization techniques are crucial to improving the generalization
pe...
Knowledge Distillation (KD) is a commonly used technique for improving t...
Knowledge distillation (KD) is an efficient framework for compressing
la...
Slot-filling and intent detection are the backbone of conversational age...
Intermediate layer knowledge distillation (KD) can improve the standard ...
Knowledge Distillation (KD) is extensively used to compress and deploy l...
Existing Natural Language Understanding (NLU) models have been shown to
...
In this work, we examine the ability of NER models to use contextual
inf...
We build a reference for the task of Open Information Extraction, on fiv...
Parallel sentence extraction is a task addressing the data sparsity prob...
Neural network approaches to Named-Entity Recognition reduce the need fo...
Parallel sentence extraction is a task addressing the data sparsity prob...