Zero-Shot Learning via Class-Conditioned Deep Generative Models

11/15/2017
by   Wenlin Wang, et al.
0

We present a deep generative model for learning to predict classes not seen at training time. Unlike most existing methods for this problem, that represent each class as a point (via a semantic embedding), we represent each seen/unseen class using a class-specific latent-space distribution, conditioned on class attributes. We use these latent-space distributions as a prior for a supervised variational autoencoder (VAE), which also facilitates learning highly discriminative feature representations for the inputs. The entire framework is learned end-to-end using only the seen-class training data. The model infers corresponding attributes of a test image by maximizing the VAE lower bound; the inferred attributes may be linked to labels not seen when training. We further extend our model to a (1) semi-supervised/transductive setting by leveraging unlabeled unseen-class data via an unsupervised learning module, and (2) few-shot learning where we also have a small number of labeled inputs from the unseen classes. We compare our model with several state-of-the-art methods through a comprehensive set of experiments on a variety of benchmark data sets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/25/2017

A Simple Exponential Family Framework for Zero-Shot Learning

We present a simple generative framework for learning to predict previou...
research
05/22/2021

Semi-Supervised Few-Shot Classification with Deep Invertible Hybrid Models

In this paper, we propose a deep invertible hybrid model which integrate...
research
12/22/2018

Can VAEs Generate Novel Examples?

An implicit goal in works on deep generative models is that such models ...
research
01/27/2019

Disentangling in Variational Autoencoders with Natural Clustering

Learning representations that disentangle the underlying factors of vari...
research
09/11/2021

Conditional Generation of Synthetic Geospatial Images from Pixel-level and Feature-level Inputs

Training robust supervised deep learning models for many geospatial appl...
research
12/08/2020

VAE-Info-cGAN: Generating Synthetic Images by Combining Pixel-level and Feature-level Geospatial Conditional Inputs

Training robust supervised deep learning models for many geospatial appl...
research
09/01/2020

Generalized Zero-Shot Learning via VAE-Conditioned Generative Flow

Generalized zero-shot learning (GZSL) aims to recognize both seen and un...

Please sign up or login with your details

Forgot password? Click here to reset