Incorporation of Human Knowledge into Data Embeddings to Improve Pattern Significance and Interpretability

09/23/2022
by   Jie Li, et al.
0

Embedding is a common technique for analyzing multi-dimensional data. However, the embedding projection cannot always form significant and interpretable visual structures that foreshadow underlying data patterns. We propose an approach that incorporates human knowledge into data embeddings to improve pattern significance and interpretability. The core idea is (1) externalizing tacit human knowledge as explicit sample labels and (2) adding a classification loss in the embedding network to encode samples' classes. The approach pulls samples of the same class with similar data features closer in the projection, leading to more compact (significant) and class-consistent (interpretable) visual structures. We give an embedding network with a customized classification loss to implement the idea and integrate the network into a visualization system to form a workflow that supports flexible class creation and pattern exploration. Patterns found on open datasets in case studies, subjects' performance in a user study, and quantitative experiment results illustrate the general usability and effectiveness of the approach.

READ FULL TEXT

page 1

page 3

page 8

page 9

research
04/24/2021

Exploring Multi-dimensional Data via Subset Embedding

Multi-dimensional data exploration is a classic research topic in visual...
research
05/28/2019

Parallax: Visualizing and Understanding the Semantics of Embedding Spaces via Algebraic Formulae

Embeddings are a fundamental component of many modern machine learning a...
research
05/09/2020

Generative Model-driven Structure Aligning Discriminative Embeddings for Transductive Zero-shot Learning

Zero-shot Learning (ZSL) is a transfer learning technique which aims at ...
research
05/29/2018

Lightly-supervised Representation Learning with Global Interpretability

We propose a lightly-supervised approach for information extraction, in ...
research
03/26/2018

HAMLET: Interpretable Human And Machine co-LEarning Technique

Efficient label acquisition processes are key to obtaining robust classi...
research
06/25/2020

Background Knowledge Injection for Interpretable Sequence Classification

Sequence classification is the supervised learning task of building mode...
research
02/09/2023

Real-Time Visual Feedback to Guide Benchmark Creation: A Human-and-Metric-in-the-Loop Workflow

Recent research has shown that language models exploit `artifacts' in be...

Please sign up or login with your details

Forgot password? Click here to reset