4.6 Article

Estimating Dynamic Treatment Regimes in Mobile Health Using V-Learning

期刊

JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION
卷 115, 期 530, 页码 692-706

出版社

AMER STATISTICAL ASSOC
DOI: 10.1080/01621459.2018.1537919

关键词

Markov decision processes; Precision medicine; Reinforcement learning; Type 1 diabetes

资金

  1. NIH [P01 CA142538, UL1 TR001111, T32 CA201159, R01 AA023187]
  2. NSF [DMS-1555141, DMS-1513579, DMS-1407732]

向作者/读者索取更多资源

The vision for precision medicine is to use individual patient characteristics to inform a personalized treatment plan that leads to the best possible healthcare for each patient. Mobile technologies have an important role to play in this vision as they offer a means to monitor a patient's health status in real-time and subsequently to deliver interventions if, when, and in the dose that they are needed. Dynamic treatment regimes formalize individualized treatment plans as sequences of decision rules, one per stage of clinical intervention, that map current patient information to a recommended treatment. However, most existing methods for estimating optimal dynamic treatment regimes are designed for a small number of fixed decision points occurring on a coarse time-scale. We propose a new reinforcement learning method for estimating an optimal treatment regime that is applicable to data collected using mobile technologies in an outpatient setting. The proposed method accommodates an indefinite time horizon and minute-by-minute decision making that are common in mobile health applications. We show that the proposed estimators are consistent and asymptotically normal under mild conditions. The proposed methods are applied to estimate an optimal dynamic treatment regime for controlling blood glucose levels in patients with type 1 diabetes.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据