DGA-Net Dynamic Gaussian Attention Network for Sentence Semantic Matching

06/09/2021
by   Kun Zhang, et al.
0

Sentence semantic matching requires an agent to determine the semantic relation between two sentences, where much recent progress has been made by the advancement of representation learning techniques and inspiration of human behaviors. Among all these methods, attention mechanism plays an essential role by selecting important parts effectively. However, current attention methods either focus on all the important parts in a static way or only select one important part at one attention step dynamically, which leaves a large space for further improvement. To this end, in this paper, we design a novel Dynamic Gaussian Attention Network (DGA-Net) to combine the advantages of current static and dynamic attention methods. More specifically, we first leverage pre-trained language model to encode the input sentences and construct semantic representations from a global perspective. Then, we develop a Dynamic Gaussian Attention (DGA) to dynamically capture the important parts and corresponding local contexts from a detailed perspective. Finally, we combine the global information and detailed local information together to decide the semantic relation of sentences comprehensively and precisely. Extensive experiments on two popular sentence semantic matching tasks demonstrate that our proposed DGA-Net is effective in improving the ability of attention mechanism.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/06/2021

LadRa-Net: Locally-Aware Dynamic Re-read Attention Net for Sentence Semantic Matching

Sentence semantic matching requires an agent to determine the semantic r...
research
12/16/2020

R^2-Net: Relation of Relation Learning Network for Sentence Semantic Matching

Sentence semantic matching is one of the fundamental tasks in natural la...
research
04/22/2018

Same Representation, Different Attentions: Shareable Sentence Representation Learning from Multiple Tasks

Distributed representation plays an important role in deep learning base...
research
12/01/2019

Not All Attention Is Needed: Gated Attention Network for Sequence Data

Although deep neural networks generally have fixed network structures, t...
research
12/19/2021

D-HAN: Dynamic News Recommendation with Hierarchical Attention Network

News recommendation is an effective information dissemination solution i...
research
04/17/2021

Context-Aware Interaction Network for Question Matching

Impressive milestones have been achieved in text matching by adopting a ...
research
10/07/2022

DABERT: Dual Attention Enhanced BERT for Semantic Matching

Transformer-based pre-trained language models such as BERT have achieved...

Please sign up or login with your details

Forgot password? Click here to reset