Hervé Jégou
Director of Sciences
Facebook AI Research
Large language models based on transformers have achieved great empirica...
The recent breakthroughs in natural language processing for model pretra...
Generative image modeling enables a wide range of applications but raise...
Lossy image compression aims to represent images in as few bits as possi...
Recent neural compression methods have been based on the popular hyperpr...
We introduce submodel co-training, a regularization method related to
co...
Image copy detection and retrieval from large databases leverage two
com...
A Vision Transformer (ViT) is a simple neural architecture amenable to s...
After their initial success in natural language processing, transformer
...
We show how to augment any convolutional network with an attention-based...
Pre-training models on large scale datasets, like ImageNet, is a standar...
We revisit watermarking techniques based on pre-trained deep networks, i...
Modern approaches for fast retrieval of similar vectors on billion-scale...
The influential Residual Networks designed by He et al. remain the
gold-...
Following their success in natural language processing, transformers hav...
We present ResMLP, an architecture built entirely upon multi-layer
perce...
In this paper, we question if self-supervised learning provides new
prop...
We propose the first general-purpose gradient-based attack against
trans...
We design a family of image classification architectures that optimize t...
Transformers have been recently adapted for large scale image classifica...
Transformers have shown outstanding results for natural language
underst...
Recently, neural networks purely based on attention were shown to addres...
This paper tackles the problem of learning a finer representation than t...
We propose a simple architecture to address unpaired image-to-image
tran...
We tackle the problem of producing compact models, maximizing their accu...
We tackle the problem of producing compact models, maximizing their accu...
This note complements the paper "Fixing the train-test resolution
discre...
We want to detect whether a particular image dataset has been used to tr...
Membership inference determines, given a sample and trained parameters o...
In this paper, we address the problem of reducing the memory footprint o...
This paper introduces a structured memory which can be easily integrated...
Transformer networks have lead to important progress in language modelin...
Data-augmentation is key to the training of neural networks for image
cl...
This paper presents a study of semi-supervised learning with large
convo...
Modern neural networks are over-parametrized. In particular, each rectif...
MultiGrain is a network architecture producing compact vector representa...
We propose a multiple-kernel local-patch descriptor based on efficient m...
Convolutional neural networks memorize part of their training data, whic...
This paper aims at learning a function mapping input vectors to an outpu...
Similarity search approaches based on graph walks have recently attained...
State-of-the-art methods for learning cross-lingual word embeddings have...
This paper aims at discovering meaningful subsets of related images from...
This paper considers the problem of inferring image labels for which onl...
Similarity search finds application in specialized database systems hand...
We consider the problem of producing compact architectures for text
clas...
We consider the design of an image representation that embeds and aggreg...
Hashing produces compact representations for documents, to perform tasks...
We propose an approximate strategy to efficiently train neural network b...
This paper considers the problem of approximate nearest neighbor search ...
This paper tackles the task of storing a large collection of vectors, su...