Interpretable Visual Reasoning via Induced Symbolic Space

11/23/2020
by   Zhonghao Wang, et al.
2

We study the problem of concept induction in visual reasoning, i.e., identifying concepts and their hierarchical relationships from question-answer pairs associated with images; and achieve an interpretable model via working on the induced symbolic concept space. To this end, we first design a new framework named object-centric compositional attention model (OCCAM) to perform the visual reasoning task with object-level visual features. Then, we come up with a method to induce concepts of objects and relations using clues from the attention patterns between objects' visual features and question words. Finally, we achieve a higher level of interpretability by imposing OCCAM on the objects represented in the induced symbolic concept space. Experiments on the CLEVR dataset demonstrate: 1) our OCCAM achieves a new state of the art without human-annotated functional programs; 2) our induced concepts are both accurate and sufficient as OCCAM achieves an on-par performance on objects represented either in visual features or in the induced symbolic concept space.

READ FULL TEXT

page 1

page 3

page 5

page 8

page 11

page 13

research
04/26/2019

The Neuro-Symbolic Concept Learner: Interpreting Scenes, Words, and Sentences From Natural Supervision

We propose the Neuro-Symbolic Concept Learner (NS-CL), a model that lear...
research
04/12/2018

Solving Bongard Problems with a Visual Language and Pragmatic Reasoning

More than 50 years ago Bongard introduced 100 visual concept learning pr...
research
10/27/2020

SIRI: Spatial Relation Induced Network For Spatial Description Resolution

Spatial Description Resolution, as a language-guided localization task, ...
research
02/04/2020

Visual Concept-Metaconcept Learning

Humans reason with concepts and metaconcepts: we recognize red and green...
research
10/01/2021

Calibrating Concepts and Operations: Towards Symbolic Reasoning on Real Images

While neural symbolic methods demonstrate impressive performance in visu...
research
04/06/2020

SHOP-VRB: A Visual Reasoning Benchmark for Object Perception

In this paper we present an approach and a benchmark for visual reasonin...
research
07/09/2023

Reasoning over the Behaviour of Objects in Video-Clips for Adverb-Type Recognition

In this work, following the intuition that adverbs describing scene-sequ...

Please sign up or login with your details

Forgot password? Click here to reset