Under mild assumptions, we investigate the structure of loss landscape o...
We propose an optimistic estimate to evaluate the best possible fitting
...
Models with nonlinear architectures/parameterizations such as deep neura...
Unraveling the general structure underlying the loss landscapes of deep
...
Substantial work indicates that the dynamics of neural networks (NNs) is...
In recent years, understanding the implicit regularization of neural net...
Understanding deep learning is increasingly emergent as it penetrates mo...
Machine learning has long been considered as a black box for predicting
...
A deep learning-based model reduction (DeePMR) method for simplifying
ch...
We prove a general Embedding Principle of loss landscape of deep neural
...
Complex design problems are common in the scientific and industrial fiel...
In this paper, we propose a model-operator-data network (MOD-Net) for so...
Understanding the structure of loss landscape of deep neural networks
(D...
It is important to study what implicit regularization is imposed on the ...
Deep neural network (DNN) usually learns the target function from low to...
Why heavily parameterized neural networks (NNs) do not overfit the data ...
A supervised learning problem is to find a function in a hypothesis func...
Recent works show an intriguing phenomenon of Frequency Principle
(F-Pri...
How neural network behaves during the training over different choices of...
This paper aims at studying the difference between Ritz-Galerkin (R-G) m...
We focus on estimating a priori generalization error of two-layer ReLU
n...
Along with fruitful applications of Deep Neural Networks (DNNs) to reali...
It remains a puzzle that why deep neural networks (DNNs), with more
para...
How different initializations and loss functions affect the learning of ...
We study the training process of Deep Neural Networks (DNNs) from the Fo...
Why deep neural networks (DNNs) capable of overfitting often generalize ...