A Kernel-Expanded Stochastic Neural Network

01/14/2022
by   Yan Sun, et al.
0

The deep neural network suffers from many fundamental issues in machine learning. For example, it often gets trapped into a local minimum in training, and its prediction uncertainty is hard to be assessed. To address these issues, we propose the so-called kernel-expanded stochastic neural network (K-StoNet) model, which incorporates support vector regression (SVR) as the first hidden layer and reformulates the neural network as a latent variable model. The former maps the input vector into an infinite dimensional feature space via a radial basis function (RBF) kernel, ensuring absence of local minima on its training loss surface. The latter breaks the high-dimensional nonconvex neural network training problem into a series of low-dimensional convex optimization problems, and enables its prediction uncertainty easily assessed. The K-StoNet can be easily trained using the imputation-regularized optimization (IRO) algorithm. Compared to traditional deep neural networks, K-StoNet possesses a theoretical guarantee to asymptotically converge to the global optimum and enables the prediction uncertainty easily assessed. The performances of the new model in training, prediction and uncertainty quantification are illustrated by simulated and real data examples.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/02/2023

Morse Neural Networks for Uncertainty Quantification

We introduce a new deep generative model useful for uncertainty quantifi...
research
12/17/2018

An Improved Deep Belief Network Model for Road Safety Analyses

Crash prediction is a critical component of road safety analyses. A wide...
research
03/02/2021

A Kernel Framework to Quantify a Model's Local Predictive Uncertainty under Data Distributional Shifts

Traditional Bayesian approaches for model uncertainty quantification rel...
research
12/24/2020

Vector-output ReLU Neural Network Problems are Copositive Programs: Convex Analysis of Two Layer Networks and Polynomial-time Algorithms

We describe the convex semi-infinite dual of the two-layer vector-output...
research
02/18/2013

Canonical dual solutions to nonconvex radial basis neural network optimization problem

Radial Basis Functions Neural Networks (RBFNNs) are tools widely used in...
research
09/07/2019

Towards Understanding the Importance of Noise in Training Neural Networks

Numerous empirical evidence has corroborated that the noise plays a cruc...
research
10/01/2021

Sparse Deep Learning: A New Framework Immune to Local Traps and Miscalibration

Deep learning has powered recent successes of artificial intelligence (A...

Please sign up or login with your details

Forgot password? Click here to reset