Neural Estimation of Statistical Divergences

10/07/2021
by   Sreejith Sreekumar, et al.
3

Statistical divergences (SDs), which quantify the dissimilarity between probability distributions, are a basic constituent of statistical inference and machine learning. A modern method for estimating those divergences relies on parametrizing an empirical variational form by a neural network (NN) and optimizing over parameter space. Such neural estimators are abundantly used in practice, but corresponding performance guarantees are partial and call for further exploration. In particular, there is a fundamental tradeoff between the two sources of error involved: approximation and empirical estimation. While the former needs the NN class to be rich and expressive, the latter relies on controlling complexity. We explore this tradeoff for an estimator based on a shallow NN by means of non-asymptotic error bounds, focusing on four popular 𝖿-divergences – Kullback-Leibler, chi-squared, squared Hellinger, and total variation. Our analysis relies on non-asymptotic function approximation theorems and tools from empirical process theory. The bounds reveal the tension between the NN size and the number of samples, and enable to characterize scaling rates thereof that ensure consistency. For compactly supported distributions, we further show that neural estimators with a slightly different NN growth-rate are near minimax rate-optimal, achieving the parametric convergence rate up to logarithmic factors.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/11/2021

Non-Asymptotic Performance Guarantees for Neural Estimation of 𝖿-Divergences

Statistical distances (SDs), which quantify the dissimilarity between pr...
research
12/19/2010

Empirical estimation of entropy functionals with confidence

This paper introduces a class of k-nearest neighbor (k-NN) estimators ca...
research
01/02/2018

A Concentration Result of Estimating Phi-Divergence using Data Dependent Partition

Estimation of the ϕ-divergence between two unknown probability distribut...
research
04/08/2023

Non-asymptotic approximations of Gaussian neural networks via second-order Poincaré inequalities

There is a growing interest on large-width asymptotic properties of Gaus...
research
07/19/2017

Rates of Uniform Consistency for k-NN Regression

We derive high-probability finite-sample uniform rates of consistency fo...
research
12/31/2020

On Gaussian Approximation for M-Estimator

This study develops a non-asymptotic Gaussian approximation theory for d...
research
02/08/2020

Extrapolation Towards Imaginary 0-Nearest Neighbour and Its Improved Convergence Rate

k-nearest neighbour (k-NN) is one of the simplest and most widely-used m...

Please sign up or login with your details

Forgot password? Click here to reset