Exploiting Inductive Bias in Transformers for Unsupervised Disentanglement of Syntax and Semantics with VAEs

05/12/2022
by   Ghazi Felhi, et al.
0

We propose a generative model for text generation, which exhibits disentangled latent representations of syntax and semantics. Contrary to previous work, this model does not need syntactic information such as constituency parses, or semantic information such as paraphrase pairs. Our model relies solely on the inductive bias found in attention-based architectures such as Transformers. In the attention of Transformers, keys handle information selection while values specify what information is conveyed. Our model, dubbed QKVAE, uses Attention in its decoder to read latent variables where one latent variable infers keys while another infers values. We run experiments on latent representations and experiments on syntax/semantics transfer which show that QKVAE displays clear signs of disentangled syntax and semantics. We also show that our model displays competitive syntax transfer capabilities when compared to supervised models and that comparable supervised models need a fairly large amount of data (more than 50K samples) to outperform it on both syntactic and semantic transfer. The code for our experiments is publicly available.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/02/2019

A Multi-Task Approach for Disentangling Syntax and Semantics in Sentence Representations

We propose a generative model for a sentence that uses two latent variab...
research
05/04/2023

Interpretable Sentence Representation with Variational Autoencoders and Attention

In this thesis, we develop methods to enhance the interpretability of re...
research
07/06/2019

Generating Sentences from Disentangled Syntactic and Semantic Spaces

Variational auto-encoders (VAEs) are widely used in natural language gen...
research
10/13/2021

Semantics-aware Attention Improves Neural Machine Translation

The integration of syntactic structures into Transformer machine transla...
research
08/27/2019

Text Modeling with Syntax-Aware Variational Autoencoders

Syntactic information contains structures and rules about how text sente...
research
12/24/2020

Disentangling semantics in language through VAEs and a certain architectural choice

We present an unsupervised method to obtain disentangled representations...
research
12/14/2022

ContraFeat: Contrasting Deep Features for Semantic Discovery

StyleGAN has shown strong potential for disentangled semantic control, t...

Please sign up or login with your details

Forgot password? Click here to reset