The Quality of the Covariance Selection Through Detection Problem and AUC Bounds

05/18/2016
by   Navid Tafaghodi Khajavi, et al.
0

We consider the problem of quantifying the quality of a model selection problem for a graphical model. We discuss this by formulating the problem as a detection problem. Model selection problems usually minimize a distance between the original distribution and the model distribution. For the special case of Gaussian distributions, the model selection problem simplifies to the covariance selection problem which is widely discussed in literature by Dempster [2] where the likelihood criterion is maximized or equivalently the Kullback-Leibler (KL) divergence is minimized to compute the model covariance matrix. While this solution is optimal for Gaussian distributions in the sense of the KL divergence, it is not optimal when compared with other information divergences and criteria such as Area Under the Curve (AUC). In this paper, we analytically compute upper and lower bounds for the AUC and discuss the quality of model selection problem using the AUC and its bounds as an accuracy measure in detection problem. We define the correlation approximation matrix (CAM) and show that analytical computation of the KL divergence, the AUC and its bounds only depend on the eigenvalues of CAM. We also show the relationship between the AUC, the KL divergence and the ROC curve by optimizing with respect to the ROC curve. In the examples provided, we pick tree structures as the simplest graphical models. We perform simulations on fully-connected graphs and compute the tree structured models by applying the widely used Chow-Liu algorithm [3]. Examples show that the quality of tree approximation models are not good in general based on information divergences, the AUC and its bounds when the number of nodes in the graphical model is large. We show both analytically and by simulations that the 1-AUC for the tree approximation model decays exponentially as the dimension of graphical model increases.

READ FULL TEXT
research
01/10/2015

On model misspecification and KL separation for Gaussian graphical models

We establish bounds on the KL divergence between two multivariate Gaussi...
research
08/10/2018

Model Approximation Using Cascade of Tree Decompositions

In this paper, we present a general, multistage framework for graphical ...
research
11/04/2020

Independent Gaussian Distributions Minimize the Kullback-Leibler (KL) Divergence from Independent Gaussian Distributions

This short note is on a property of the Kullback-Leibler (KL) divergence...
research
03/11/2020

Gaussian Graphical Model exploration and selection in high dimension low sample size setting

Gaussian Graphical Models (GGM) are often used to describe the condition...
research
11/09/2020

Near-Optimal Learning of Tree-Structured Distributions by Chow-Liu

We provide finite sample guarantees for the classical Chow-Liu algorithm...
research
08/20/2018

Use Of Vapnik-Chervonenkis Dimension in Model Selection

In this dissertation, I derive a new method to estimate the Vapnik-Cherv...
research
11/03/2018

Stochastic Neighbor Embedding under f-divergences

The t-distributed Stochastic Neighbor Embedding (t-SNE) is a powerful an...

Please sign up or login with your details

Forgot password? Click here to reset