4.6 Article

Testing the Plasticity of Reinforcement Learning-based Systems

出版社

ASSOC COMPUTING MACHINERY
DOI: 10.1145/3511701

关键词

Software testing; reinforcement learning; empirical software engineering

资金

  1. ERC [787703]
  2. European Research Council (ERC) [787703] Funding Source: European Research Council (ERC)

向作者/读者索取更多资源

This article presents a method to test the plasticity of reinforcement learning-based systems. It quantifies the adaptation and anti-regression capabilities of the system by computing its adaptation frontier in a changed environment. Visualizing the results provides developers with crucial information for deciding whether to enable online learning or not.
The dataset available for pre-release training of a machine-learning based system is often not representative of all possible execution contexts that the system will encounter in the field. Reinforcement Learning (RL) is a prominent approach among those that support continual learning, i.e., learning continually in the field, in the post-release phase. No study has so far investigated any method to test the plasticity of RL-based systems, i.e., their capability to adapt to an execution context that may deviate from the training one. We propose an approach to test the plasticity of RL-based systems. The output of our approach is a quantification of the adaptation and anti-regression capabilities of the system, obtained by computing the adaptation frontier of the system in a changed environment. We visualize such frontier as an adaptation/anti-regression heatmap in two dimensions, or as a clustered projection when more than two dimensions are involved. In this way, we provide developers with information on the amount of changes that can be accommodated by the continual learning component of the system, which is key to decide if online, in-the-field learning can be safely enabled or not.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据