4.7 Article

Deep reinforcement learning driven inspection and maintenance planning under incomplete information and constraints

Journal

RELIABILITY ENGINEERING & SYSTEM SAFETY
Volume 212, Issue -, Pages -

Publisher

ELSEVIER SCI LTD
DOI: 10.1016/j.ress.2021.107551

Keywords

Inspection and maintenance planning; System risk and reliability; Constrained stochastic optimization; Partially observable Markov decision processes; Deep reinforcement learning; Decentralized multi-agent control

Funding

  1. U.S. National Science Foundation [1751941]
  2. Center for Integrated Asset Management for Multimodal Transportation Infrastructure Systems (CIAMTIS), 2018 U.S. DOT Region 3 University Center
  3. Div Of Civil, Mechanical, & Manufact Inn
  4. Directorate For Engineering [1751941] Funding Source: National Science Foundation

Ask authors/readers for more resources

Determining inspection and maintenance policies to minimize long-term risks and costs in deteriorating engineering environments is a complex optimization problem. Major computational challenges include the curse of dimensionality, curse of history, presence of state uncertainties, and presence of constraints. These challenges are addressed through a joint framework of constrained Partially Observable Markov Decision Processes (POMDP) and multi-agent Deep Reinforcement Learning (DRL).
Determination of inspection and maintenance policies for minimizing long-term risks and costs in deteriorating engineering environments constitutes a complex optimization problem. Major computational challenges include the (i) curse of dimensionality, due to exponential scaling of state/action set cardinalities with the number of components; (ii) curse of history, related to exponentially growing decision-trees with the number of decisionsteps; (iii) presence of state uncertainties, induced by inherent environment stochasticity and variability of inspection/monitoring measurements; (iv) presence of constraints, pertaining to stochastic long-term limitations, due to resource scarcity and other infeasible/undesirable system responses. In this work, these challenges are addressed within a joint framework of constrained Partially Observable Markov Decision Processes (POMDP) and multi-agent Deep Reinforcement Learning (DRL). POMDPs optimally tackle (ii)-(iii), combining stochastic dynamic programming with Bayesian inference principles. Multi-agent DRL addresses (i), through deep function parametrizations and decentralized control assumptions. Challenge (iv) is herein handled through proper state augmentation and Lagrangian relaxation, with emphasis on life-cycle risk-based constraints and budget limitations. The underlying algorithmic steps are provided, and the proposed framework is found to outperform well-established policy baselines and facilitate adept prescription of inspection and intervention actions, in cases where decisions must be made in the most resource- and risk-aware manner.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available