It is well known that randomly initialized, push-forward, fully connected neural networks weakly converge to isotropic Gaussian processes in the limit where the width of all layers goes to infinity. In this paper, we propose to use the angular power spectrum of the limiting fields to characterize the complexity of the network architecture. In particular, we define sequences of random variables associated with the angular power spectrum and provide a full characterization of the network complexity in terms of the asymptotic distribution of these sequences as the depth diverges. On this basis, we classify neural networks as low-disorder, sparse, or high-disorder; we show how this classification highlights a number of distinct features for standard activation functions and, in particular, sparsity properties of ReLU networks. Our theoretical results are also validated by numerical simulations.

Di Lillo, S., Marinucci, D., Salvi, M., Vigogna, S. (2025). Spectral Complexity of Deep Neural Networks. SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 7(3), 1154-1183 [10.1137/24m1675746].

Spectral Complexity of Deep Neural Networks

Di Lillo, Simmaco;Marinucci, Domenico;Salvi, Michele;Vigogna, Stefano
2025-01-01

Abstract

It is well known that randomly initialized, push-forward, fully connected neural networks weakly converge to isotropic Gaussian processes in the limit where the width of all layers goes to infinity. In this paper, we propose to use the angular power spectrum of the limiting fields to characterize the complexity of the network architecture. In particular, we define sequences of random variables associated with the angular power spectrum and provide a full characterization of the network complexity in terms of the asymptotic distribution of these sequences as the depth diverges. On this basis, we classify neural networks as low-disorder, sparse, or high-disorder; we show how this classification highlights a number of distinct features for standard activation functions and, in particular, sparsity properties of ReLU networks. Our theoretical results are also validated by numerical simulations.
2025
Pubblicato
Rilevanza internazionale
Articolo
Esperti anonimi
Settore MATH-03/B - Probabilità e statistica matematica
English
Con Impact Factor ISI
angular power spectrum
compositional kernels
deep learning
Gaussian processes
isotropic random fields
neural networks
Di Lillo, S., Marinucci, D., Salvi, M., Vigogna, S. (2025). Spectral Complexity of Deep Neural Networks. SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 7(3), 1154-1183 [10.1137/24m1675746].
Di Lillo, S; Marinucci, D; Salvi, M; Vigogna, S
Articolo su rivista
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2108/444685
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
social impact