Absorbing Phase Transitions in Artificial Deep Neural Networks
https://arxiv.org/abs/2307.02284

To summarize, we believe that the this work places the order-to-chaos transition in the initialized artificial deep neural networks in the broader context of absorbing phase transitions, & serves as the first step toward the systematic comparison between natural/biological & artificial neural networks.
...

#NeuralNetworks #MeanFieldTheory #SignalPropagation #PhaseTransitions #backpropagation #feedforward

Universal Scaling Laws of Absorbing Phase Transitions in Artificial Deep Neural Networks

We demonstrate that conventional artificial deep neural networks operating near the phase boundary of the signal propagation dynamics, also known as the edge of chaos, exhibit universal scaling laws of absorbing phase transitions in non-equilibrium statistical mechanics. We exploit the fully deterministic nature of the propagation dynamics to elucidate an analogy between a signal collapse in the neural networks and an absorbing state (a state that the system can enter but cannot escape from). Our numerical results indicate that the multilayer perceptrons and the convolutional neural networks belong to the mean-field and the directed percolation universality classes, respectively. Also, the finite-size scaling is successfully applied, suggesting a potential connection to the depth-width trade-off in deep learning. Furthermore, our analysis of the training dynamics under the gradient descent reveals that hyperparameter tuning to the phase boundary is necessary but insufficient for achieving optimal generalization in deep networks. Remarkably, nonuniversal metric factors associated with the scaling laws are shown to play a significant role in concretizing the above observations. These findings highlight the usefulness of the notion of criticality for analyzing the behavior of artificial deep neural networks and offer new insights toward a unified understanding of the essential relationship between criticality and intelligence.

arXiv.org

...
We foresee some interesting directions for future work. One of the most natural directions is to extend the present analysis to the backpropagation and to analyze the training dynamics.

Mean-field theory: https://en.wikipedia.org/wiki/Mean-field_theory

Absorbing Markov chain: https://en.wikipedia.org/wiki/Absorbing_Markov_chain
An absorbing state is a state that, once entered, cannot be left.

Discrete phase-type distribution: https://en.wikipedia.org/wiki/Discrete_phase-type_distribution

Mean-field theory - Wikipedia