Wasserstein Dependency Measure for Representation Learning

03/28/2019
by   Sherjil Ozair, et al.
18

Mutual information maximization has emerged as a powerful learning objective for unsupervised representation learning obtaining state-of-the-art performance in applications such as object recognition, speech recognition, and reinforcement learning. However, such approaches are fundamentally limited since a tight lower bound of mutual information requires sample size exponential in the mutual information. This limits the applicability of these approaches for prediction tasks with high mutual information, such as in video understanding or reinforcement learning. In these settings, such techniques are prone to overfit, both in theory and in practice, and capture only a few of the relevant factors of variation. This leads to incomplete representations that are not optimal for downstream tasks. In this work, we empirically demonstrate that mutual information-based representation learning approaches do fail to learn complete representations on a number of designed and real-world tasks. To mitigate these problems we introduce the Wasserstein dependency measure, which learns more complete representations by using the Wasserstein distance instead of the KL divergence in the mutual information estimator. We show that a practical approximation to this theoretically motivated solution, constructed using Lipschitz constraint techniques from the GAN literature, achieves substantially improved results on tasks where incomplete representations are a major challenge.

READ FULL TEXT
research
03/08/2021

Multimodal Representation Learning via Maximization of Local Mutual Information

We propose and demonstrate a representation learning approach by maximiz...
research
11/04/2022

Unsupervised Visual Representation Learning via Mutual Information Regularized Assignment

This paper proposes Mutual Information Regularized Assignment (MIRA), a ...
research
03/13/2020

DHOG: Deep Hierarchical Object Grouping

Recently, a number of competitive methods have tackled unsupervised repr...
research
01/28/2023

Mutual Wasserstein Discrepancy Minimization for Sequential Recommendation

Self-supervised sequential recommendation significantly improves recomme...
research
06/11/2021

Robust Representation Learning via Perceptual Similarity Metrics

A fundamental challenge in artificial intelligence is learning useful re...
research
07/04/2022

Representation Learning with Information Theory for COVID-19 Detection

Successful data representation is a fundamental factor in machine learni...
research
06/14/2021

Which Mutual-Information Representation Learning Objectives are Sufficient for Control?

Mutual information maximization provides an appealing formalism for lear...

Please sign up or login with your details

Forgot password? Click here to reset