VGSE: Visually-Grounded Semantic Embeddings for Zero-Shot Learning

03/20/2022
by   Wenjia Xu, et al.
5

Human-annotated attributes serve as powerful semantic embeddings in zero-shot learning. However, their annotation process is labor-intensive and needs expert supervision. Current unsupervised semantic embeddings, i.e., word embeddings, enable knowledge transfer between classes. However, word embeddings do not always reflect visual similarities and result in inferior zero-shot performance. We propose to discover semantic embeddings containing discriminative visual properties for zero-shot learning, without requiring any human annotation. Our model visually divides a set of images from seen classes into clusters of local image regions according to their visual similarity, and further imposes their class discrimination and semantic relatedness. To associate these clusters with previously unseen classes, we use external knowledge, e.g., word embeddings and propose a novel class relation discovery module. Through quantitative and qualitative evaluation, we demonstrate that our model discovers semantic embeddings that model the visual properties of both seen and unseen classes. Furthermore, we demonstrate on three benchmarks that our visually-grounded semantic embeddings further improve performance over word embeddings across various ZSL models by a large margin.

READ FULL TEXT

page 3

page 8

research
07/18/2017

Visually Aligned Word Embeddings for Improving Zero-shot Learning

Zero-shot learning (ZSL) highly depends on a good semantic embedding to ...
research
10/06/2020

Using Sentences as Semantic Representations in Large Scale Zero-Shot Learning

Zero-shot learning aims to recognize instances of unseen classes, for wh...
research
09/21/2022

I2DFormer: Learning Image to Document Attention for Zero-Shot Image Classification

Despite the tremendous progress in zero-shot learning(ZSL), the majority...
research
01/03/2022

Semantically Grounded Visual Embeddings for Zero-Shot Learning

Zero-shot learning methods rely on fixed visual and semantic embeddings,...
research
12/11/2020

Improving Zero Shot Learning Baselines with Commonsense Knowledge

Zero shot learning – the problem of training and testing on a completely...
research
09/21/2020

Visual-Semantic Embedding Model Informed by Structured Knowledge

We propose a novel approach to improve a visual-semantic embedding model...
research
06/05/2017

Learning Structured Semantic Embeddings for Visual Recognition

Numerous embedding models have been recently explored to incorporate sem...

Please sign up or login with your details

Forgot password? Click here to reset