Link the head to the "beak": Zero Shot Learning from Noisy Text Description at Part Precision

09/04/2017
by   Mohamed Elhoseiny, et al.
0

In this paper, we study learning visual classifiers from unstructured text descriptions at part precision with no training images. We propose a learning framework that is able to connect text terms to its relevant parts and suppress connections to non-visual text terms without any part-text annotations. For instance, this learning process enables terms like "beak" to be sparsely linked to the visual representation of parts like head, while reduces the effect of non-visual terms like "migrate" on classifier prediction. Images are encoded by a part-based CNN that detect bird parts and learn part-specific representation. Part-based visual classifiers are predicted from text descriptions of unseen visual classifiers to facilitate classification without training images (also known as zero-shot recognition). We performed our experiments on CUBirds 2011 dataset and improves the state-of-the-art text-based zero-shot recognition results from 34.7% to 43.6%. We also created large scale benchmarks on North American Bird Images augmented with text descriptions, where we also show that our approach outperforms existing methods. Our code, data, and models are publically available.

READ FULL TEXT

page 4

page 8

research
10/07/2020

ZEST: Zero-shot Learning from Text Descriptions using Textual Similarity and Visual Summarization

We study the problem of recognizing visual entities from the textual des...
research
10/27/2022

Text2Model: Model Induction for Zero-shot Generalization Using Task Descriptions

We study the problem of generating a training-free task-dependent visual...
research
06/29/2015

Tell and Predict: Kernel Classifier Prediction for Unseen Visual Classes from Unstructured Text Descriptions

In this paper we propose a framework for predicting kernelized classifie...
research
04/20/2022

Unsupervised Ranking and Aggregation of Label Descriptions for Zero-Shot Classifiers

Zero-shot text classifiers based on label descriptions embed an input te...
research
12/31/2015

Write a Classifier: Predicting Visual Classifiers from Unstructured Text

People typically learn through exposure to visual concepts associated wi...
research
11/10/2022

Zero-shot Visual Commonsense Immorality Prediction

Artificial intelligence is currently powering diverse real-world applica...
research
09/15/2017

Zero-Shot Learning to Manage a Large Number of Place-Specific Compressive Change Classifiers

With recent progress in large-scale map maintenance and long-term map le...

Please sign up or login with your details

Forgot password? Click here to reset