4.8 Article

Learning Representations for Neural Network-Based Classification Using the Information Bottleneck Principle

Journal

Publisher

IEEE COMPUTER SOC
DOI: 10.1109/TPAMI.2019.2909031

Keywords

Training; Task analysis; Robustness; Cost function; Neurons; Neural networks; Deep learning; information bottleneck; representation learning; regularization; classification; neural networks; stochastic neural networks

Funding

  1. German Federal Ministry of Education and Research
  2. Erwin Schrodinger Fellowship of the Austrian Science Fund [J 3765]
  3. Austrian COMET Program -Competence Centers for Excellent Technologies -under Austrian Federal Ministry of Transport, Innovation and Technology
  4. Austrian Federal Ministry of Digital and Economic Affairs
  5. State of Styria

Ask authors/readers for more resources

In this theory paper, we investigate training deep neural networks (DNNs) for classification via minimizing the information bottleneck (IB) functional. We show that the resulting optimization problem suffers from two severe issues: First, for deterministic DNNs, either the IB functional is infinite for almost all values of network parameters, making the optimization problem ill-posed, or it is piecewise constant, hence not admitting gradient-based optimization methods. Second, the invariance of the IB functional under bijections prevents it from capturing properties of the learned representation that are desirable for classification, such as robustness and simplicity. We argue that these issues are partly resolved for stochastic DNNs, DNNs that include a (hard or soft) decision rule, or by replacing the IB functional with related, but more well-behaved cost functions. We conclude that recent successes reported about training DNNs using the IB framework must be attributed to such solutions. As a side effect, our results indicate limitations of the IB framework for the analysis of DNNs. We also note that rather than trying to repair the inherent problems in the IB functional, a better approach may be to design regularizers on latent representation enforcing the desired properties directly.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available