4.7 Article

Strategy Synthesis for POMDPs in Robot Planning via Game-Based Abstractions

期刊

IEEE TRANSACTIONS ON AUTOMATIC CONTROL
卷 66, 期 3, 页码 1040-1054

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TAC.2020.2990140

关键词

Planning; Games; Probabilistic logic; Markov processes; Safety; Model checking; Probability distribution; Partially observable Markov decision process (POMDP); probabilistic model checking; probabilistic two-player game (PG); robot navigation

资金

  1. CDZ Project CAP [GZ 1023]
  2. German Research Foundation (DFG) as part of the Cluster of Excellence BrainLinks/BrainTools [EXC 1086]
  3. ONR [N000141612051]
  4. NASA [80NSSC19K0209]
  5. DARPA [W911NF-16-1-0001]
  6. German Research Foundation (DFG) [RTG 2236]
  7. U.S. Department of Defense (DOD) [N000141612051] Funding Source: U.S. Department of Defense (DOD)

向作者/读者索取更多资源

This research focuses on synthesis problems with constraints in POMDPs and proposes an abstraction refinement framework to transform a POMDP model into a probabilistic two-player game, allowing for determination of optimal strategies through efficient verification and synthesis tools. This method advances the state of the art in solving planning problems in partially observable environments with safety guarantees.
We study synthesis problems with constraints in partially observable Markov decision processes (POMDPs), where the objective is to compute a strategy for an agent that is guaranteed to satisfy certain safety and performance specifications. Verification and strategy synthesis for POMDPs are, however, computationally intractable in general. We alleviate this difficulty by focusing on planning applications and exploiting typical structural properties of such scenarios; for instance, we assume that the agent has the ability to observe its own position inside an environment. We propose an abstraction refinement framework, which turns such a POMDP model into a (fully observable) probabilistic two-player game (PG). For the obtained PGs, efficient verification and synthesis tools allow to determine strategies with optimal safety and performance measures, which approximate optimal schedulers on the POMDP. If the approximation is too coarse to satisfy the given specifications, a refinement scheme improves the computed strategies. As a running example, we use planning problems where an agent moves inside an environment with randomly moving obstacles and restricted observability. We demonstrate that the proposed method advances the state of the art by solving problems several orders of magnitude larger than those that can be handled by existing POMDP solvers. Furthermore, this method gives guarantees on safety constraints, which is not supported by the majority of the existing solvers.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据