One Simple Trick to Fix Your Bayesian Neural Network

07/26/2022
by   Piotr Tempczyk, et al.
0

One of the most popular estimation methods in Bayesian neural networks (BNN) is mean-field variational inference (MFVI). In this work, we show that neural networks with ReLU activation function induce posteriors, that are hard to fit with MFVI. We provide a theoretical justification for this phenomenon, study it empirically, and report the results of a series of experiments to investigate the effect of activation function on the calibration of BNNs. We find that using Leaky ReLU activations leads to more Gaussian-like weight posteriors and achieves a lower expected calibration error (ECE) than its ReLU-based counterpart.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/09/2022

On the Activation Function Dependence of the Spectral Bias of Neural Networks

Neural networks are universal function approximators which are known to ...
research
12/06/2018

Singular Values for ReLU Layers

Despite their prevalence in neural networks we still lack a thorough the...
research
02/23/2022

Wide Mean-Field Bayesian Neural Networks Ignore the Data

Bayesian neural networks (BNNs) combine the expressive power of deep lea...
research
11/24/2017

Invariance of Weight Distributions in Rectified MLPs

An interesting approach to analyzing and developing tools for neural net...
research
03/22/2020

Dynamic ReLU

Rectified linear units (ReLU) are commonly used in deep neural networks....
research
07/06/2016

A Modified Activation Function with Improved Run-Times For Neural Networks

In this paper we present a modified version of the Hyperbolic Tangent Ac...
research
06/29/2023

A Quantitative Functional Central Limit Theorem for Shallow Neural Networks

We prove a Quantitative Functional Central Limit Theorem for one-hidden-...

Please sign up or login with your details

Forgot password? Click here to reset