Uncertainty Estimation of Transformers' Predictions via Topological Analysis of the Attention Matrices

08/22/2023
by   Elizaveta Kostenok, et al.
0

Determining the degree of confidence of deep learning model in its prediction is an open problem in the field of natural language processing. Most of the classical methods for uncertainty estimation are quite weak for text classification models. We set the task of obtaining an uncertainty estimate for neural networks based on the Transformer architecture. A key feature of such mo-dels is the attention mechanism, which supports the information flow between the hidden representations of tokens in the neural network. We explore the formed relationships between internal representations using Topological Data Analysis methods and utilize them to predict model's confidence. In this paper, we propose a method for uncertainty estimation based on the topological properties of the attention mechanism and compare it with classical methods. As a result, the proposed algorithm surpasses the existing methods in quality and opens up a new area of application of the attention mechanism, but requires the selection of topological features.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/05/2022

Betti numbers of attention graphs is all you really need

We apply methods of topological analysis to the attention graphs, calcul...
research
12/19/2022

Uncovering the Origins of Instability in Dynamical Systems: How Attention Mechanism Can Help?

The behavior of the network and its stability are governed by both dynam...
research
02/18/2020

Text Classification with Lexicon from PreAttention Mechanism

A comprehensive and high-quality lexicon plays a crucial role in traditi...
research
10/31/2022

QNet: A Quantum-native Sequence Encoder Architecture

This work investigates how current quantum computers can improve the per...
research
05/24/2018

Uncertainty-Aware Attention for Reliable Interpretation and Prediction

Attention mechanism is effective in both focusing the deep learning mode...
research
11/14/2022

Do Neural Networks Trained with Topological Features Learn Different Internal Representations?

There is a growing body of work that leverages features extracted via to...
research
03/14/2022

Simplicial Attention Neural Networks

The aim of this work is to introduce simplicial attention networks (SANs...

Please sign up or login with your details

Forgot password? Click here to reset