4.7 Article

Understanding Neural Networks and Individual Neuron Importance via Information-Ordered Cumulative Ablation

Journal

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TNNLS.2021.3088685

Keywords

Neurons; Artificial neural networks; Mutual information; Entropy; Training; Task analysis; Biological neural networks; Ablation analysis; deep learning; information theory; neural networks (NNs)

Funding

  1. German Federal Ministry of Education and Research through the Framework of the Alexander von Humboldt-Professorship
  2. Austrian Science Fund
  3. iDev40 Project through the Erwin Schrodinger Fellowship [J 3765]
  4. ECSEL Joint Undertaking (JU) [783163]
  5. European Union's Horizon 2020 Research and Innovation Programme

Ask authors/readers for more resources

In this study, three information-theoretic quantities were used to analyze the behavior of trained neural networks, revealing that class selectivity is not a reliable indicator for classification performance. However, when examining individual layers, mutual information and class selectivity show a positive correlation with classification performance.
In this work, we investigate the use of three information-theoretic quantities--entropy, mutual information with the class variable, and a class selectivity measure based on Kullback-Leibler (KL) divergence--to understand and study the behavior of already trained fully connected feedforward neural networks (NNs). We analyze the connection between these information-theoretic quantities and classification performance on the test set by cumulatively ablating neurons in networks trained on MNIST, FashionMNIST, and CIFAR-10. Our results parallel those recently published by Morcos et al., indicating that class selectivity is not a good indicator for classification performance. However, looking at individual layers separately, both mutual information and class selectivity are positively correlated with classification performance, at least for networks with ReLU activation functions. We provide explanations for this phenomenon and conclude that it is ill-advised to compare the proposed information-theoretic quantities across layers. Furthermore, we show that cumulative ablation of neurons with ascending or descending information-theoretic quantities can be used to formulate hypotheses regarding the joint behavior of multiple neurons, such as redundancy and synergy, with comparably low computational cost. We also draw connections to the information bottleneck theory for NNs.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available