Does a Neural Network Really Encode Symbolic Concept?

02/25/2023
by   Mingjie Li, et al.
0

Recently, a series of studies have tried to extract interactions between input variables modeled by a DNN and define such interactions as concepts encoded by the DNN. However, strictly speaking, there still lacks a solid guarantee whether such interactions indeed represent meaningful concepts. Therefore, in this paper, we examine the trustworthiness of interaction concepts from four perspectives. Extensive empirical studies have verified that a well-trained DNN usually encodes sparse, transferable, and discriminative concepts, which is partially aligned with human intuition.

READ FULL TEXT

page 13

page 14

page 16

page 17

page 18

research
04/26/2023

Technical Note: Defining and Quantifying AND-OR Interactions for Faithful and Concise Explanation of DNNs

In this technical note, we aim to explain a deep neural network (DNN) by...
research
04/03/2023

Can the Inference Logic of Large Language Models be Disentangled into Symbolic Concepts?

In this paper, we explain the inference logic of large language models (...
research
05/03/2023

Where We Have Arrived in Proving the Emergence of Sparse Symbolic Concepts in AI Models

This paper aims to prove the emergence of symbolic concepts in well-trai...
research
06/21/2021

A Game-Theoretic Taxonomy of Visual Concepts in DNNs

In this paper, we rethink how a DNN encodes visual concepts of different...
research
07/31/2021

A Hypothesis for the Aesthetic Appreciation in Neural Networks

This paper proposes a hypothesis for the aesthetic appreciation that aes...
research
02/25/2023

Bayesian Neural Networks Tend to Ignore Complex and Sensitive Concepts

In this paper, we focus on mean-field variational Bayesian Neural Networ...
research
01/03/2014

ConceptVision: A Flexible Scene Classification Framework

We introduce ConceptVision, a method that aims for high accuracy in cate...

Please sign up or login with your details

Forgot password? Click here to reset