On the curvature of the loss landscape

07/10/2023
by   Alison Pouplin, et al.
0

One of the main challenges in modern deep learning is to understand why such over-parameterized models perform so well when trained on finite data. A way to analyze this generalization concept is through the properties of the associated loss landscape. In this work, we consider the loss landscape as an embedded Riemannian manifold and show that the differential geometric properties of the manifold can be used when analyzing the generalization abilities of a deep net. In particular, we focus on the scalar curvature, which can be computed analytically for our manifold, and show connections to several settings that potentially imply generalization.

READ FULL TEXT

page 1

page 4

research
07/22/2023

The instabilities of large learning rate training: a loss landscape view

Modern neural networks are undeniably successful. Numerous works study h...
research
03/22/2018

Curvature of Hypergraphs via Multi-Marginal Optimal Transport

We introduce a novel definition of curvature for hypergraphs, a natural ...
research
08/22/2018

Statistical Neurodynamics of Deep Networks: Geometry of Signal Spaces

Statistical neurodynamics studies macroscopic behaviors of randomly conn...
research
09/19/2022

Neural Collapse with Normalized Features: A Geometric Analysis over the Riemannian Manifold

When training overparameterized deep networks for classification tasks, ...
research
12/20/2019

MLRG Deep Curvature

We present MLRG Deep Curvature suite, a PyTorch-based, open-source packa...
research
04/03/2023

Charting the Topography of the Neural Network Landscape with Thermal-Like Noise

The training of neural networks is a complex, high-dimensional, non-conv...
research
06/18/2019

Information matrices and generalization

This work revisits the use of information criteria to characterize the g...

Please sign up or login with your details

Forgot password? Click here to reset