Shaping Visual Representations with Language for Few-shot Classification

11/06/2019
by   Jesse Mu, et al.
0

Language is designed to convey useful information about the world, thus serving as a scaffold for efficient human learning. How can we let language guide representation learning in machine learning models? We explore this question in the setting of few-shot visual classification, proposing models which learn to perform visual classification while jointly predicting natural language task descriptions at train time. At test time, with no language available, we find that these language-influenced visual representations are more generalizable, compared to meta-learning baselines and approaches that explicitly use language as a bottleneck for classification.

READ FULL TEXT
research
07/07/2022

Improving Few-Shot Image Classification Using Machine- and User-Generated Natural Language Descriptions

Humans can obtain the knowledge of novel visual concepts from language d...
research
07/10/2023

Leveraging Multiple Descriptive Features for Robust Few-shot Image Learning

Modern image classification is based upon directly predicting model clas...
research
10/17/2022

Meta-Learning via Classifier(-free) Guidance

State-of-the-art meta-learning techniques do not optimize for zero-shot ...
research
03/27/2020

Detection and Description of Change in Visual Streams

This paper presents a framework for the analysis of changes in visual st...
research
02/09/2022

Predicting Human Similarity Judgments Using Large Language Models

Similarity judgments provide a well-established method for accessing men...
research
05/25/2020

Incidental Supervision: Moving beyond Supervised Learning

Machine Learning and Inference methods have become ubiquitous in our att...
research
01/13/2023

Infusing Commonsense World Models with Graph Knowledge

While language models have become more capable of producing compelling l...

Please sign up or login with your details

Forgot password? Click here to reset