4.6 Article

Hierarchical Interpretation of Neural Text Classification

Journal

COMPUTATIONAL LINGUISTICS
Volume 48, Issue 4, Pages 987-1020

Publisher

MIT PRESS
DOI: 10.1162/coli_a_00459

Keywords

-

Funding

  1. UK Engineering and Physical Sciences Research Council [EP/T017112/1, EP/V048597/1, EP/X019063/1]
  2. University of Warwick
  3. Chinese Scholarship Council
  4. UK Research and Innovation [EP/V020579/1]

Ask authors/readers for more resources

In recent years, there has been increasing interest in developing interpretable models in Natural Language Processing (NLP). However, it is difficult to accurately explain model decisions by words or phrases when neural models in NLP compose word semantics hierarchically. This article proposes a novel Hierarchical Interpretable Neural Text classifier, called HINT, which generates explanations of model predictions in the form of label-associated topics. Experimental results show that HINT achieves comparable text classification results and provides better interpretations than other interpretable neural text classifiers.
Recent years have witnessed increasing interest in developing interpretable models in Natural Language Processing (NLP). Most existing models aim at identifying input features such as words or phrases important for model predictions. Neural models developed in NLP, however, often compose word semantics in a hierarchical manner. As such, interpretation by words or phrases only cannot faithfully explain model decisions in text classification. This article proposes a novel Hierarchical Interpretable Neural Text classifier, called HINT, which can automatically generate explanations of model predictions in the form of label-associated topics in a hierarchical manner. Model interpretation is no longer at the word level, but built on topics as the basic semantic unit. Experimental results on both review datasets and news datasets show that our proposed approach achieves text classification results on par with existing state-of-the-art text classifiers, and generates interpretations more faithful to model predictions and better understood by humans than other interpretable neural text classifiers.(1)

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available