Tell and Predict: Kernel Classifier Prediction for Unseen Visual Classes from Unstructured Text Descriptions

06/29/2015
by   Mohamed Elhoseiny, et al.
0

In this paper we propose a framework for predicting kernelized classifiers in the visual domain for categories with no training images where the knowledge comes from textual description about these categories. Through our optimization framework, the proposed approach is capable of embedding the class-level knowledge from the text domain as kernel classifiers in the visual domain. We also proposed a distributional semantic kernel between text descriptions which is shown to be effective in our setting. The proposed framework is not restricted to textual descriptions, and can also be applied to other forms knowledge representations. Our approach was applied for the challenging task of zero-shot learning of fine-grained categories from text descriptions of these categories.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/31/2015

Write a Classifier: Predicting Visual Classifiers from Unstructured Text

People typically learn through exposure to visual concepts associated wi...
research
09/04/2017

Link the head to the "beak": Zero Shot Learning from Noisy Text Description at Part Precision

In this paper, we study learning visual classifiers from unstructured te...
research
10/07/2020

ZEST: Zero-shot Learning from Text Descriptions using Textual Similarity and Visual Summarization

We study the problem of recognizing visual entities from the textual des...
research
11/18/2020

A Multi-class Approach – Building a Visual Classifier based on Textual Descriptions using Zero-Shot Learning

Machine Learning (ML) techniques for image classification routinely requ...
research
06/01/2015

Predicting Deep Zero-Shot Convolutional Neural Networks using Textual Descriptions

One of the main challenges in Zero-Shot Learning of visual categories is...
research
04/10/2022

Explaining Deep Convolutional Neural Networks via Latent Visual-Semantic Filter Attention

Interpretability is an important property for visual models as it helps ...
research
07/29/2023

Sat2Cap: Mapping Fine-Grained Textual Descriptions from Satellite Images

We propose a novel weakly supervised approach for creating maps using fr...

Please sign up or login with your details

Forgot password? Click here to reset