On the curvature of the loss landscape
https://arxiv.org/abs/2307.04719

A main challenge in modern deep learning is to understand why such over-parameterized models perform so well when trained on finite data ... we consider the loss landscape as an embedded Riemannian manifold ... we focus on the scalar curvature, which can be computed analytically for our manifold ...

Manifolds: https://en.wikipedia.org/wiki/Manifold
...

#MachineLearning #NeuralNetworks #manifolds #parametrization #DeepLearning #LossFunctions

On the curvature of the loss landscape

One of the main challenges in modern deep learning is to understand why such over-parameterized models perform so well when trained on finite data. A way to analyze this generalization concept is through the properties of the associated loss landscape. In this work, we consider the loss landscape as an embedded Riemannian manifold and show that the differential geometric properties of the manifold can be used when analyzing the generalization abilities of a deep net. In particular, we focus on the scalar curvature, which can be computed analytically for our manifold, and show connections to several settings that potentially imply generalization.

arXiv.org

...
Addendum

Dimensionality reduction
https://en.wikipedia.org/wiki/Machine_learning#Dimensionality_reduction

Dimensionality reduction reduces the number of random variables under consideration by obtaining a set of principal variables. PCA changes higher-dimensional data (e.g. 3D) to a smaller space (e.g. 2D)

The manifold hypothesis proposes that high-dimensional data sets lie along low-dimensional manifolds.
Many dimensionality reduction techniques make this assumption, leading to the area of manifold learning and manifold regularization

Machine learning - Wikipedia

...
Addendae (cont'd)

Manifold hypothesis
https://en.wikipedia.org/wiki/Manifold_hypothesis

Many high-dimensional data sets (requiring many variables) in the real world actually lie along low-dimensional latent manifolds in that high-dimensional space (described by a smaller number of variables).

This principle may underpin the effectiveness of ML algorithms in describing high-dimensional data sets by considering a few common features.

#ManifoldHypothesis #manifolds #MachineLearning #DimensionalityReduction

Manifold hypothesis - Wikipedia