4.7 Article

Spatial relation learning for explainable image classification and annotation in critical applications

Journal

ARTIFICIAL INTELLIGENCE
Volume 292, Issue -, Pages -

Publisher

ELSEVIER
DOI: 10.1016/j.artint.2020.103434

Keywords

Explainable artificial intelligence; Relation learning; Fuzzy logic

Ask authors/readers for more resources

This article proposes an approach for explainable classification and annotation of images in high-stake applications, based on a transparent model and interpretable fuzzy relations. By setting expert knowledge in advance and extracting relevant relations using a fuzzy frequent itemset mining algorithm, rules for classification and constraints for annotation are successfully built. Experimental results demonstrate the approach's ability to successfully perform the target task and generate consistent and convincing explanations.
With the recent successes of black-box models in Artificial Intelligence (AI) and the growing interactions between humans and AIs, explainability issues have risen. In this article, in the context of high-stake applications, we propose an approach for explainable classification and annotation of images. It is based on a transparent model, whose reasoning is accessible and human understandable, and on interpretable fuzzy relations that enable to express the vagueness of natural language. The knowledge about relations is set beforehand by an expert and thus training instances do not need to be annotated. The most relevant relations are extracted using a fuzzy frequent itemset mining algorithm in order to build rules, for classification, and constraints, for annotation. We also present two heuristics that make the process of evaluating relations faster. Since the strengths of our approach are the transparency of the model and the interpretability of the relations, an explanation in natural language can be generated. Supported by experimental results, we show that, given a segmentation of the input, our approach is able to successfully perform the target task and generate explanations that were judged as consistent and convincing by a set of participants. (C) 2020 Elsevier B.V. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available