Link Prediction with Mutual Attention for Text-Attributed Networks

02/28/2019
by   Robin Brochier, et al.
0

In this extended abstract, we present an algorithm that learns a similarity measure between documents from the network topology of a structured corpus. We leverage the Scaled Dot-Product Attention, a recently proposed attention mechanism, to design a mutual attention mechanism between pairs of documents. To train its parameters, we use the network links as supervision. We provide preliminary experiment results with a citation dataset on two prediction tasks, demonstrating the capacity of our model to learn a meaningful textual similarity.

READ FULL TEXT

page 1

page 2

research
09/15/2017

Query-based Attention CNN for Text Similarity Map

In this paper, we introduce Query-based Attention CNN(QACNN) for Text Si...
research
03/15/2018

Aggregated Sparse Attention for Steering Angle Prediction

In this paper, we apply the attention mechanism to autonomous driving fo...
research
09/30/2020

Learning Hard Retrieval Cross Attention for Transformer

The Transformer translation model that based on the multi-head attention...
research
03/25/2021

A comparative analysis of local network similarity measurements: application to author citation networks

Understanding the evolution of paper and author citations is of paramoun...
research
04/18/2022

BSAL: A Framework of Bi-component Structure and Attribute Learning for Link Prediction

Given the ubiquitous existence of graph-structured data, learning the re...
research
06/11/2020

Attention improves concentration when learning node embeddings

We consider the problem of predicting edges in a graph from node attribu...
research
06/05/2023

Infusing Lattice Symmetry Priors in Attention Mechanisms for Sample-Efficient Abstract Geometric Reasoning

The Abstraction and Reasoning Corpus (ARC) (Chollet, 2019) and its most ...

Please sign up or login with your details

Forgot password? Click here to reset