Most current approaches for protecting privacy in machine learning (ML)
...
The legality of training language models (LMs) on copyrighted or otherwi...
An emerging method to cheaply improve a weaker language model is to fine...
Instruction-tuned LMs such as ChatGPT, FLAN, and InstructGPT are finetun...
Image diffusion models such as DALL-E 2, Imagen, and Stable Diffusion ha...
The internet contains a wealth of knowledge – from the birthdays of
hist...
We present the Berkeley Crossword Solver, a state-of-the-art approach fo...
Code is seldom written in a single left-to-right pass and is instead
rep...
Past work has shown that large language models are susceptible to privac...
To create models that are robust across a wide range of test inputs, tra...
Prompting language models (LMs) with training examples and task descript...
Language models (LMs) must be both safe and equitable to be responsibly
...
GPT-3 can perform numerous tasks when provided a natural language prompt...
It has become common to publish large (billion parameter) language model...
The remarkable success of pretrained language models has motivated the s...
Adversarial attacks alter NLP model predictions by perturbing test-time
...
Gradient-based analysis methods, such as saliency map visualizations and...
In this work, we provide an industry research view for approaching the
d...
We consider an adversary looking to steal or attack a black-box machine
...
Although pretrained Transformers such as BERT achieve high accuracy on
i...
Standard test sets for supervised learning evaluate in-distribution
gene...
Since hardware resources are limited, the objective of training deep lea...
Neural NLP models are increasingly accurate but are imperfect and
opaque...
The ability to understand and work with numbers (numeracy) is critical f...
Adversarial examples highlight model vulnerabilities and are useful for
...
Adversarial examples highlight model vulnerabilities and are useful for
...
Multi-hop reading comprehension (RC) questions are challenging because t...
Recent work establishes dataset difficulty and removes annotation artifa...
Current methods to interpret deep learning models by generating saliency...
Local model interpretation methods explain individual predictions by
ass...
Modern natural language processing systems have been touted as approachi...
Exposing the weaknesses of neural models is crucial for improving their
...