4.7 Review

Explainable AI for earth observation: A review including societal and regulatory perspectives

Publisher

ELSEVIER
DOI: 10.1016/j.jag.2022.102869

Keywords

Earth observation; Remote sensing; Machine learning; Explainable artificial intelligence; Ethics; Regulations

Categories

Funding

  1. Dutch Research Council (NWO) [18091]

Ask authors/readers for more resources

This paper reviews examples of explainable machine learning and explainable artificial intelligence in the field of Earth Observation, classifying the methods and identifying limitations. The findings indicate a lack of clarity, with explanations often targeting domain experts and lacking testing for usefulness to the intended audience.
Artificial intelligence and machine learning are ubiquitous in the domain of Earth Observation (EO) and Remote Sensing. Congruent to their success in the domain of computer vision, they have proven to obtain high accuracies for EO applications. Yet experts of EO should also consider the weaknesses of complex, machine-learning models before adopting them for specific applications. One such weakness is the lack of explainability of complex deep learning models. This paper reviews published examples of explainable ML or explainable AI in the field of Earth Observation. Explainability methods are classified as: intrinsic versus post-hoc, model-specific versus model -agnostic, and global versus local explanations and examples of each type are provided. This paper also iden-tifies key explainability requirements identified the social sciences and upcoming regulatory recommendations from UNESCO Ethics of Artificial Intelligence and requirements from the EU draft Artificial Intelligence Act and analyzes whether these limitations are sufficiently addressed in the field of EO.The findings indicate that there is a lack of clarity regarding which models can be considered interpretable or not. EO applications often utilize Random Forests as an interpretable benchmark algorithm to compare to complex deep-learning models even though social sciences clearly argue that large Random Forests cannot be considered as such. Secondly, most explanations target domain experts and not possible users of the algorithm, regulatory bodies, or those who might be affected by an algorithm's decisions. Finally, publications tend to simply provide explanations without testing the usefulness of the explanation by the intended audience. In light of these societal and regulatory considerations, a framework is provided to guide the selection of an appropriate machine learning algorithm based on the availability of simpler algorithms with a high predictive accuracy as well as the purpose and intended audience of the explanation.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available