We investigate the predictability of large language model (LLM) capabili...
Large Language Models (LLMs) have exhibited an impressive ability to per...
Recent work suggests that transformer models are capable of multi-task
l...
Distilling state-of-the-art transformer models into lightweight student
...
Current NLP models are predominantly trained through a pretrain-then-fin...
Neural networks are prone to learning spurious correlations from biased
...
Pre-trained text-to-text transformers achieve impressive performance acr...
Closed-book question-answering (QA) is a challenging task that requires ...
For decades, Internet protocols have been specified using natural langua...
Advances in extractive machine reading comprehension (MRC) rely heavily ...
Deep neural networks usually require massive labeled data, which restric...
Deep neural networks usually require massive labeled data, which restric...
In recent years there is surge of interest in applying distant supervisi...