4.7 Article

Reinforcement learning-based expanded personalized diabetes treatment recommendation using South Korean electronic health records

Journal

EXPERT SYSTEMS WITH APPLICATIONS
Volume 206, Issue -, Pages -

Publisher

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.eswa.2022.117932

Keywords

Reinforcement learning; Precision medicine; Decision making; Data -driven optimization; Electronic health records

Funding

  1. National Research Foundation of Korea (NRF) - Korean government (MSIP) [NRF- 2018R1D1A1A02046351]
  2. Electronics and Telecommunications Research Institute (ETRI) - ICT R & D program of MSIT/IITP [2014-3-00123]

Ask authors/readers for more resources

In this study, the researchers aim to develop a reinforcement learning-based treatment recommendation model using electronic health records of South Korean patients. By considering various details, the proposed contextual bandits model offers a practical solution to address clinical challenges in prescribing diabetes medication.
Currently, electronic medical records are becoming more accessible to a growing number of researchers seeking to develop personalized healthcare recommendations to aid physicians in making better clinical decisions and treating patients. As a result, clinical decision research has become more focused on data-driven optimization. In this study, we analyze Korean patients' electronic health records-including medical history, medications, laboratory tests, and more information-shared by the national health insurance system. We aim to develop a reinforcement learning-based expanded treatment recommendation model using the health records of South Korean citizens to assist physicians. This study is significant in that expert and intelligent systems harmoniously solve the problem that directly addresses many clinical challenges in prescribing proper diabetes medication when assessing the physical state of diabetes patients. Reinforcement learning is a mechanism for determining how agents should behave in a given environment to maximize a cumulative reward. The basic model for a reinforcement learning design environment is the Markov decision process (MDP) model. Although it is effective and easy to use, the MDP model is limited by dimensionality, i.e., many details about the patients cannot be considered when building the model. To address this issue, we applied a contextual bandits approach to create a more practical model that can expand states and actions by considering several details that are crucial for pa-tients with diabetes. Finally, we validated the performance of the proposed contextual bandits model by comparing it with existing reinforcement-learning algorithms.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available