Most current approaches for protecting privacy in machine learning (ML)
...
Because state-of-the-art language models are expensive to train, most
pr...
Large language models are now tuned to align with the goals of their
cre...
If machine learning models were to achieve superhuman abilities at vario...
Decision-based evasion attacks repeatedly query a black-box classifier t...
It is becoming increasingly imperative to design robust ML defenses. How...
Deep learning models are often trained on distributed, webscale datasets...
Auditing mechanisms for differential privacy use probabilistic means to
...
Image diffusion models such as DALL-E 2, Imagen, and Stable Diffusion ha...
The performance of differentially private machine learning can be booste...
Studying data memorization in neural language models helps us understand...
Stable Diffusion is a recent open-source image generation model comparab...
Property inference attacks allow an adversary to extract global properti...
Hundreds of defenses have been proposed to make deep neural networks rob...
Machine learning models trained on private datasets have been shown to l...
We introduce a new class of attacks on machine learning models. We show ...
Large language models (LMs) have been shown to memorize parts of their
t...
Natural language reflects our private lives and identities, making its
p...
Modern neural language models widely used in tasks across NLP risk memor...
Differentially Private (DP) learning has seen limited success for buildi...
NeuraCrypt (Yara et al. arXiv 2021) is an algorithm that converts a sens...
Making classifiers robust to adversarial examples is hard. Thus, many
de...
Data poisoning has been proposed as a compelling defense against facial
...
We consider the privacy-preserving machine learning (ML) setting where t...
It has become common to publish large (billion parameter) language model...
We demonstrate that differentially private machine learning has not yet
...
Membership inference attacks are one of the simplest forms of privacy le...
Adaptive attacks have (rightfully) become the de facto standard for
eval...
Adversarial examples are malicious inputs crafted to induce
misclassific...
Federated learning (FL) is a machine learning setting where many clients...
Incentive mechanisms are central to the functionality of permissionless
...
Defenses against adversarial examples, such as adversarial training, are...
Adversarial examples are malicious inputs crafted to cause a model to
mi...
SentiNet is a novel detection framework for physical attacks on neural
n...
Perceptual ad-blocking is a novel approach that uses visual cues to dete...
Deep neural networks (DNNs) are vulnerable to adversarial
examples-malic...
As Machine Learning (ML) gets applied to security-critical or sensitive
...
Deep learning has proven to be a powerful tool for computer vision and h...
Machine learning models are vulnerable to adversarial examples, inputs
m...
Adversarial examples are maliciously perturbed inputs designed to mislea...
Machine learning (ML) models may be deemed confidential due to their
sen...