Pathological Spectra of the Fisher Information Metric and Its Variants in Deep Neural Networks
Autor: | Shun-ichi Amari, Shotaro Akaho, Ryo Karakida |
---|---|
Rok vydání: | 2021 |
Předmět: |
FOS: Computer and information sciences
Hessian matrix Computer Science - Machine Learning Cognitive Neuroscience Feature vector FOS: Physical sciences Machine Learning (stat.ML) 02 engineering and technology 01 natural sciences Machine Learning (cs.LG) Machine Learning 010104 statistics & probability symbols.namesake Arts and Humanities (miscellaneous) Statistics - Machine Learning 0202 electrical engineering electronic engineering information engineering Metric tensor 0101 mathematics Fisher information Mathematics 020206 networking & telecommunications Disordered Systems and Neural Networks (cond-mat.dis-nn) Condensed Matter - Disordered Systems and Neural Networks Kernel (statistics) Metric (mathematics) Softmax function symbols Neural Networks Computer Algorithm Fisher information metric |
Zdroj: | Neural Computation. 33:2274-2307 |
ISSN: | 1530-888X 0899-7667 |
DOI: | 10.1162/neco_a_01411 |
Popis: | The Fisher information matrix (FIM) plays an essential role in statistics and machine learning as a Riemannian metric tensor or a component of the Hessian matrix of loss functions. Focusing on the FIM and its variants in deep neural networks (DNNs), we reveal their characteristic scale dependence on the network width, depth and sample size when the network has random weights and is sufficiently wide. This study covers two widely-used FIMs for regression with linear output and for classification with softmax output. Both FIMs asymptotically show pathological eigenvalue spectra in the sense that a small number of eigenvalues become large outliers depending the width or sample size while the others are much smaller. It implies that the local shape of the parameter space or loss landscape is very sharp in a few specific directions while almost flat in the other directions. In particular, the softmax output disperses the outliers and makes a tail of the eigenvalue density spread from the bulk. We also show that pathological spectra appear in other variants of FIMs: one is the neural tangent kernel; another is a metric for the input signal and feature space that arises from feedforward signal propagation. Thus, we provide a unified perspective on the FIM and its variants that will lead to more quantitative understanding of learning in large-scale DNNs. 23 pages, 7 figures; v2: minor improvements, Section 3.4 added |
Databáze: | OpenAIRE |
Externí odkaz: |