Autor: |
Baglioni, P., Giambagli, L., Vezzani, A., Burioni, R., Rotondo, P., Pacelli, R. |
Rok vydání: |
2024 |
Předmět: |
|
Druh dokumentu: |
Working Paper |
Popis: |
Finite-width one hidden layer networks with multiple neurons in the readout layer display non-trivial output-output correlations that vanish in the lazy-training infinite-width limit. In this manuscript we leverage recent progress in the proportional limit of Bayesian deep learning (that is the limit where the size of the training set $P$ and the width of the hidden layers $N$ are taken to infinity keeping their ratio $\alpha = P/N$ finite) to rationalize this empirical evidence. In particular, we show that output-output correlations in finite fully-connected networks are taken into account by a kernel shape renormalization of the infinite-width NNGP kernel, which naturally arises in the proportional limit. We perform accurate numerical experiments both to assess the predictive power of the Bayesian framework in terms of generalization, and to quantify output-output correlations in finite-width networks. By quantitatively matching our predictions with the observed correlations, we provide additional evidence that kernel shape renormalization is instrumental to explain the phenomenology observed in finite Bayesian one hidden layer networks. |
Databáze: |
arXiv |
Externí odkaz: |
|