Aligning language models (LMs) with preferences is an important problem ...
AI tasks encompass a wide range of domains and fields. While numerous AI...
Through prompting, large-scale pre-trained models have become more expre...
When learning task-oriented dialogue (ToD) agents, reinforcement learnin...
Unsupervised clustering under domain shift (UCDS) studies how to transfe...
Retriever-reader models achieve competitive performance across many diff...
In offline model-based reinforcement learning (offline MBRL), we learn a...
Offline reinforcement learning (RL) extends the paradigm of classical RL...
Active learning, which effectively collects informative unlabeled data f...
Generating images from natural language instructions is an intriguing ye...
The neural attention mechanism has been incorporated into deep neural
ne...
Training NLP systems typically assumes access to annotated data that has...
Attention-based neural networks have achieved state-of-the-art results o...
We study calibration in question answering, estimating whether model
cor...
Dropout has been demonstrated as a simple and effective module to not on...
We study estimating inherent human disagreement (annotation label
distri...
Attention modules, as simple and effective tools, have not only enabled ...