4.2 Article

Event-driven temporal models for explanations-ETeMoX: explaining reinforcement learning

Journal

SOFTWARE AND SYSTEMS MODELING
Volume 21, Issue 3, Pages 1091-1113

Publisher

SPRINGER HEIDELBERG
DOI: 10.1007/s10270-021-00952-4

Keywords

Temporal models; Complex event processing; Artificial intelligence; Explainable reinforcement learning; Event-driven monitoring

Funding

  1. The Leverhulme Trust Fellowship QuantUn: quantification of uncertainty using Bayesian surprises [RF-2019-548/9]
  2. EPSRC Research Project [EP/T017627/1]
  3. Royal Society of Edinburgh project A Reinforcement Learning Based Resource Management System for Long Term Care for Elderly People [961_Yang]
  4. Spanish Ministry of Science and Innovation
  5. European Regional Development Funds under project FAME [RTI2018-093608-B-C33]
  6. Research Plan from the University of Cadiz
  7. Grupo Ener-getico de Puerto Real S.A. under project GANGES [IRTP03_UCA]

Ask authors/readers for more resources

The increasing autonomy in modern software systems, especially in the context of Reinforcement Learning, raises concerns about the transparency of decision-making criteria, requiring solutions for explainability and trustworthiness in AI systems.
Modern software systems are increasingly expected to show higher degrees of autonomy and self-management to cope with uncertain and diverse situations. As a consequence, autonomous systems can exhibit unexpected and surprising behaviours. This is exacerbated due to the ubiquity and complexity of Artificial Intelligence (AI)-based systems. This is the case of Reinforcement Learning (RL), where autonomous agents learn through trial-and-error how to find good solutions to a problem. Thus, the underlying decision-making criteria may become opaque to users that interact with the system and who may require explanations about the system's reasoning. Available work for eXplainable Reinforcement Learning (XRL) offers different trade-offs: e.g. for runtime explanations, the approaches are model-specific or can only analyse results after-the-fact. Different from these approaches, this paper aims to provide an online model-agnostic approach for XRL towards trustworthy and understandable AI. We present ETeMoX, an architecture based on temporal models to keep track of the decision-making processes of RL systems. In cases where the resources are limited (e.g. storage capacity or time to response), the architecture also integrates complex event processing, an event-driven approach, for detecting matches to event patterns that need to be stored, instead of keeping the entire history. The approach is applied to a mobile communications case study that uses RL for its decision-making. In order to test the generalisability of our approach, three variants of the underlying RL algorithms are used: Q-Learning, SARSA and DQN. The encouraging results show that using the proposed configurable architecture, RL developers are able to obtain explanations about the evolution of a metric, relationships between metrics, and were able to track situations of interest happening over time windows.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.2
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available