On Mutual Information Maximization for Representation Learning

07/31/2019
by   Michael Tschannen, et al.
0

Many recent methods for unsupervised or self-supervised representation learning train feature extractors by maximizing an estimate of the mutual information (MI) between different views of the data. This comes with several immediate problems: For example, MI is notoriously hard to estimate, and using it as an objective for representation learning may lead to highly entangled representations due to its invariance under arbitrary invertible transformations. Nevertheless, these methods have been repeatedly shown to excel in practice. In this paper we argue, and provide empirical evidence, that the success of these methods might be only loosely attributed to the properties of MI, and that they strongly depend on the inductive bias in both the choice of feature extractor architectures and the parametrization of the employed MI estimators. Finally, we establish a connection to deep metric learning and argue that this interpretation may be a plausible explanation for the success of the recently introduced methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/14/2021

Maximizing Mutual Information Across Feature and Topology Views for Learning Graph Representations

Recently, maximizing mutual information has emerged as a powerful method...
research
10/18/2019

A Mutual Information Maximization Perspective of Language Representation Learning

We show state-of-the-art word representation learning methods maximize a...
research
05/03/2020

Mutual Information Gradient Estimation for Representation Learning

Mutual Information (MI) plays an important role in representation learni...
research
01/19/2023

DiME: Maximizing Mutual Information by a Difference of Matrix-Based Entropies

We introduce an information-theoretic quantity with similar properties t...
research
06/03/2019

Learning Representations by Maximizing Mutual Information Across Views

We propose an approach to self-supervised representation learning based ...
research
06/14/2021

Which Mutual-Information Representation Learning Objectives are Sufficient for Control?

Mutual information maximization provides an appealing formalism for lear...
research
07/07/2020

Hierarchical and Unsupervised Graph Representation Learning with Loukas's Coarsening

We propose a novel algorithm for unsupervised graph representation learn...

Please sign up or login with your details

Forgot password? Click here to reset