Does a Neural Network Really Encode Symbolic Concept?

02/25/2023
by   Mingjie Li, et al.
0

Recently, a series of studies have tried to extract interactions between input variables modeled by a DNN and define such interactions as concepts encoded by the DNN. However, strictly speaking, there still lacks a solid guarantee whether such interactions indeed represent meaningful concepts. Therefore, in this paper, we examine the trustworthiness of interaction concepts from four perspectives. Extensive empirical studies have verified that a well-trained DNN usually encodes sparse, transferable, and discriminative concepts, which is partially aligned with human intuition.

READ FULL TEXT
research
04/26/2023

Technical Note: Defining and Quantifying AND-OR Interactions for Faithful and Concise Explanation of DNNs

In this technical note, we aim to explain a deep neural network (DNN) by...
research
04/03/2023

Can the Inference Logic of Large Language Models be Disentangled into Symbolic Concepts?

In this paper, we explain the inference logic of large language models (...
research
05/03/2023

Where We Have Arrived in Proving the Emergence of Sparse Symbolic Concepts in AI Models

This paper aims to prove the emergence of symbolic concepts in well-trai...
research
06/21/2021

A Game-Theoretic Taxonomy of Visual Concepts in DNNs

In this paper, we rethink how a DNN encodes visual concepts of different...
research
07/31/2021

A Hypothesis for the Aesthetic Appreciation in Neural Networks

This paper proposes a hypothesis for the aesthetic appreciation that aes...
research
02/25/2023

Bayesian Neural Networks Tend to Ignore Complex and Sensitive Concepts

In this paper, we focus on mean-field variational Bayesian Neural Networ...
research
01/03/2014

ConceptVision: A Flexible Scene Classification Framework

We introduce ConceptVision, a method that aims for high accuracy in cate...

Please sign up or login with your details

Forgot password? Click here to reset