Recent work has shown that fine-tuning large pre-trained language models...
Large language models (LLMs) have exhibited remarkable capabilities in
l...
Mixture of Experts layers (MoEs) enable efficient scaling of language mo...
Do language models have beliefs about the world? Dennett (1995) famously...
Current abstractive summarization systems outperform their extractive
co...
Natural language (NL) explanations of model predictions are gaining
popu...
While research on explaining predictions of open-domain QA systems (ODQA...
State-of-the-art Machine Reading Comprehension (MRC) models for Open-dom...
We present ELQ, a fast end-to-end entity linking model for questions, wh...
We introduce a very deep and light-weight transformer, DeLighT, that del...
Interactive programming with interleaved code snippet cells and natural
...
Programmers typically organize executable source code using high-level c...
Source code is rarely written in isolation. It depends significantly on ...
We propose a context-dependent model to map utterances within an interac...
We present an approach to rapidly and easily build natural language
inte...
Sequence-to-sequence models have shown strong performance across a broad...