Why Shallow Networks Struggle with Approximating and Learning High Frequency: A Numerical Study

06/29/2023
by   Shijun Zhang, et al.
0

In this work, a comprehensive numerical study involving analysis and experiments shows why a two-layer neural network has difficulties handling high frequencies in approximation and learning when machine precision and computation cost are important factors in real practice. In particular, the following fundamental computational issues are investigated: (1) the best accuracy one can achieve given a finite machine precision, (2) the computation cost to achieve a given accuracy, and (3) stability with respect to perturbations. The key to the study is the spectral analysis of the corresponding Gram matrix of the activation functions which also shows how the properties of the activation function play a role in the picture.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/16/2023

Data-aware customization of activation functions reduces neural network error

Activation functions play critical roles in neural networks, yet current...
research
05/25/2023

Embeddings between Barron spaces with higher order activation functions

The approximation properties of infinitely wide shallow neural networks ...
research
02/07/2020

Ill-Posedness and Optimization Geometry for Nonlinear Neural Network Training

In this work we analyze the role nonlinear activation functions play at ...
research
05/24/2022

Taming the sign problem of explicitly antisymmetrized neural networks via rough activation functions

Explicit antisymmetrization of a two-layer neural network is a potential...
research
10/27/2018

A Methodology for Automatic Selection of Activation Functions to Design Hybrid Deep Neural Networks

Activation functions influence behavior and performance of DNNs. Nonline...
research
08/30/2023

A spectrum adaptive kernel polynomial method

The kernel polynomial method (KPM) is a powerful numerical method for ap...
research
04/12/2023

Numerical differentiation by the polynomial-exponential basis

Our objective is to calculate the derivatives of data corrupted by noise...

Please sign up or login with your details

Forgot password? Click here to reset