Nearest Class-Center Simplification through Intermediate Layers

Autor: Ben-Shaul, Ido, Dekel, Shai
Rok vydání: 2022
Předmět:
Druh dokumentu: Working Paper
Popis: Recent advances in theoretical Deep Learning have introduced geometric properties that occur during training, past the Interpolation Threshold -- where the training error reaches zero. We inquire into the phenomena coined Neural Collapse in the intermediate layers of the networks, and emphasize the innerworkings of Nearest Class-Center Mismatch inside the deepnet. We further show that these processes occur both in vision and language model architectures. Lastly, we propose a Stochastic Variability-Simplification Loss (SVSL) that encourages better geometrical features in intermediate layers, and improves both train metrics and generalization.
Databáze: arXiv