Skip to content

Official code for our paper "Intermediate Hidden Layers of Classification Neural Networks" in ICML 2023's Workshop in High-Dimensional Learning Dynamics.

Notifications You must be signed in to change notification settings

emreonal11/Neural-Collapse-in-Intermediate-Hidden-Layers-of-Classification-Neural-Networks

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 

Repository files navigation

Neural-Collapse-in-Intermediate-Hidden-Layers-of-Classification-Neural-Networks

Official code for our paper "Intermediate Hidden Layers of Classification Neural Networks" in ICML 2023's Workshop in High-Dimensional Learning Dynamics.

Neural collapse ($\mathcal{NC}$) gives a precise description of the representations of classes in the final hidden layer of classification neural networks. This description provides insights into how these networks learn features and generalize well when trained past zero training error. However, to date, $\mathcal{NC}$ has only been studied in the final layer of these networks. In the present paper, we provide the first comprehensive empirical analysis of the emergence of $\mathcal{NC}$ in the intermediate hidden layers of these classifiers. We examine a variety of network architectures, activations, and datasets, and demonstrate that some degree of $\mathcal{NC}$ emerges in most of the intermediate hidden layers of the network, where the degree of collapse in any given layer is typically positively correlated with the depth of that layer in the neural network. Moreover, we remark that: (1) almost all of the reduction in intra-class variance in the samples occurs in the shallower layers of the networks, (2) the angular separation between class means increases consistently with hidden layer depth, and (3) simple datasets require only the shallower layers of the networks to fully learn them, whereas more difficult ones require the entire network. Ultimately, these results provide granular insights into the structural propagation of features through classification neural networks.

Please refer to our paper for more information: https://arxiv.org/abs/2308.02760

About

Official code for our paper "Intermediate Hidden Layers of Classification Neural Networks" in ICML 2023's Workshop in High-Dimensional Learning Dynamics.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages