4.7 Article

Online Altitude Control and Scheduling Policy for Minimizing AoI in UAV-Assisted IoT Wireless Networks

期刊

IEEE TRANSACTIONS ON MOBILE COMPUTING
卷 21, 期 7, 页码 2493-2505

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TMC.2020.3042925

关键词

Optimization; Relays; Internet of Things; Trajectory; Real-time systems; Unmanned aerial vehicles; Dynamic scheduling; Mobile relays; age of information; scheduling policy; UAV altitude control; proximal policy optimization algorithm; unknown channel conditions

资金

  1. Concordia University
  2. FQRNT

向作者/读者索取更多资源

This article discusses UAV-assisted IoT networks and proposes an online model-free deep reinforcement learning approach to optimize the scheduling policy for transmitting information from low-resource IoT devices to base stations. By dynamically adjusting the deployment altitude of UAVs, the proposed method minimizes the age of information for IoT devices. Extensive simulations demonstrate the effectiveness of the proposed design.
This article considers unmanned aerial vehicle (UAV) assisted Internet of Things (IoT) networks, where low resource IoT devices periodically sample a stochastic process and need to upload more recent information to a Base Station (BS). Among the myriad of applications, there is a need for timely delivery of data (for example, status-updates) before the data becomes outdated and loses its value. Since transmission capabilities of IoT devices are limited, it may not always be feasible to transmit over one hop transmission to the BS. To address this challenge, UAVs with virtual queues are deployed as middle layer between IoT devices and the BS to relay recent information over unreliable channels. In the absence of channel conditions, the optimal online scheduling policy is investigated as well as dynamic UAV altitude control that maintains a fresh status of information at the BS. The objective of this paper is to minimize the Expected Weighted Sum Age of Information (EWSA) for IoT devices. First, the problem is formulated as an optimization problem that is however generally hard to solve. Second, an online model free Deep Reinforcement Learning (DRL) is proposed, where the deployed UAV obtains instantaneous channel state information (CSI) in real time along with any adjustment to its deployment altitude. Third, we formulate the online problem as a Markov Decision Process (MDP) and Proximal Policy Optimization (PPO) algorithm, which is a highly stable state-of-the-art DRL algorithm, is leveraged to solve the formulated problem. Finally, extensive simulations are conducted to verify findings and comprehensive comparisons with other baseline approaches are provided to demonstrate the effectiveness of the proposed design.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据