Training large language models to follow instructions makes them perform...
Every major technical invention resurfaces the dual-use dilemma – the ne...
Analysts and scientists are interested in querying streams of video, aud...
Generative AI, in particular text-based "foundation models" (large model...
We propose a methodology for planting watermarks in text from an
autoreg...
Language models (LMs) are increasingly being used in open-ended contexts...
Existing foundation models are trained on copyrighted material. Deployin...
Despite increasingly fluent, relevant, and coherent language generation,...
Models trained on one set of domains often suffer performance drops on u...
Recent advances in instruction-following large language models (LLMs) ha...
Recent work has identified noisy and misannotated data as a core cause o...
Task-oriented dialogue systems often assist users with personal or
confi...
Language models (LMs) are becoming the foundation for almost all major
l...
Over the past few years, AI methods of generating images have been incre...
Machine learning models are now able to convert user-written text
descri...
Likelihood, although useful as a training loss, is a poor search objecti...
As ML models have increased in capabilities and accuracy, so has the
com...
We systematically study the calibration of classifiers trained with
diff...
Large pretrained models can be privately fine-tuned to achieve performan...
While a broad range of techniques have been proposed to tackle distribut...
Model-based, reference-free evaluation metrics have been proposed as a f...
As machine learning models are deployed ever more broadly, it becomes
in...
Modern language models can generate high-quality short texts. However, t...
Whose labels should a machine learning (ML) algorithm learn to emulate? ...
Importance weighting is a classic technique to handle distribution shift...
Machine learning systems deployed in the wild are often trained on a sou...
Differentially Private (DP) learning has seen limited success for buildi...
In conversation, uptake happens when a speaker builds on the contributio...
We study how masking and predicting tokens in an unsupervised fashion ca...
Distributionally robust optimization (DRO) provides a framework for trai...
We introduce GEM, a living benchmark for natural language Generation (NL...
Unstructured data is now commonly queried by using target deep neural
ne...
While modern large-scale datasets often consist of heterogeneous
subpopu...
The reliability of machine learning systems critically assumes that the
...
Neural language models are usually trained to match the distributional
p...
Due to the falling costs of data acquisition and storage, researchers an...
Modeling how individuals evolve over time is a fundamental problem in th...