Contrastive estimation reveals topic posterior information to linear models

03/04/2020
by   Christopher Tosh, et al.
6

Contrastive learning is an approach to representation learning that utilizes naturally occurring similar and dissimilar pairs of data points to find useful embeddings of data. In the context of document classification under topic modeling assumptions, we prove that contrastive learning is capable of recovering a representation of documents that reveals their underlying topic posterior information to linear models. We apply this procedure in a semi-supervised setup and demonstrate empirically that linear classifiers with these representations perform well in document classification tasks with very few training examples.

READ FULL TEXT
research
08/24/2020

Contrastive learning, multi-view redundancy, and linear models

Self-supervised learning is an empirically successful approach to unsupe...
research
03/26/2021

Unsupervised Document Embedding via Contrastive Augmentation

We present a contrasting learning approach with data augmentation techni...
research
06/24/2015

Efficient Learning for Undirected Topic Models

Replicated Softmax model, a well-known undirected topic model, is powerf...
research
05/25/2023

Efficient Document Embeddings via Self-Contrastive Bregman Divergence Learning

Learning quality document embeddings is a fundamental problem in natural...
research
06/09/2022

I'm Me, We're Us, and I'm Us: Tri-directional Contrastive Learning on Hypergraphs

Although machine learning on hypergraphs has attracted considerable atte...
research
12/30/2009

MedLDA: A General Framework of Maximum Margin Supervised Topic Models

Supervised topic models utilize document's side information for discover...
research
02/27/2022

UCTopic: Unsupervised Contrastive Learning for Phrase Representations and Topic Mining

High-quality phrase representations are essential to finding topics and ...

Please sign up or login with your details

Forgot password? Click here to reset