Consistent Sparse Deep Learning: Theory and Computation

02/25/2021
by   Yan Sun, et al.
0

Deep learning has been the engine powering many successes of data science. However, the deep neural network (DNN), as the basic model of deep learning, is often excessively over-parameterized, causing many difficulties in training, prediction and interpretation. We propose a frequentist-like method for learning sparse DNNs and justify its consistency under the Bayesian framework: the proposed method could learn a sparse DNN with at most O(n/log(n)) connections and nice theoretical guarantees such as posterior consistency, variable selection consistency and asymptotically optimal generalization bounds. In particular, we establish posterior consistency for the sparse DNN with a mixture Gaussian prior, show that the structure of the sparse DNN can be consistently determined using a Laplace approximation-based marginal posterior inclusion probability approach, and use Bayesian evidence to elicit sparse DNNs learned by an optimization method such as stochastic gradient descent in multiple runs with different initializations. The proposed method is computationally more efficient than standard Bayesian methods for large-scale sparse DNNs. The numerical results indicate that the proposed method can perform very well for large-scale network compression and high-dimensional nonlinear variable selection, both advancing interpretable machine learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/14/2023

Empirical Bayes inference in sparse high-dimensional generalized linear models

High-dimensional linear models have been extensively studied in the rece...
research
02/16/2022

APPLADE: Adjustable Plug-and-play Audio Declipper Combining DNN with Sparse Optimization

In this paper, we propose an audio declipping method that takes advantag...
research
11/19/2020

Application of Deep Learning-based Interpolation Methods to Nearshore Bathymetry

Nearshore bathymetry, the topography of the ocean floor in coastal zones...
research
04/15/2022

Towards a Unified Framework for Uncertainty-aware Nonlinear Variable Selection with Theoretical Guarantees

We develop a simple and unified framework for nonlinear variable selecti...
research
11/14/2022

Inconsistency identification in network meta-analysis via stochastic search variable selection

The reliability of the results of network meta-analysis (NMA) lies in th...
research
10/01/2021

Sparse Deep Learning: A New Framework Immune to Local Traps and Miscalibration

Deep learning has powered recent successes of artificial intelligence (A...
research
12/12/2019

Variable Selection Consistency of Gaussian Process Regression

Bayesian nonparametric regression under a rescaled Gaussian process prio...

Please sign up or login with your details

Forgot password? Click here to reset