4.7 Article

Trajectory Design and Power Control for Multi-UAV Assisted Wireless Networks: A Machine Learning Approach

Journal

IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY
Volume 68, Issue 8, Pages 7957-7969

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TVT.2019.2920284

Keywords

Multi-agent Q-learning; power control; trajectory design; Twitter; unmanned aerial vehicle (UAV)

Funding

  1. Engineering and Physical Sciences Research Council [EP/Noo4558/1, EP/PO34284/1]
  2. COALESCE of the Royal Society's Global Challenges Research Fund Grant
  3. European Research Council's Advanced Fellow Grant QuantCom
  4. EPSRC [EP/L010550/1, EP/J015520/1, EP/P003990/1, EP/N004558/1] Funding Source: UKRI

Ask authors/readers for more resources

A novel framework is proposed for the trajectory design of multiple unmanned aerial vehicles (UAVs) based on the prediction of users' mobility information. The problem of joint trajectory design and power control is formulated for maximizing the instantaneous sum transmit rate while satisfying the rate requirement of users. In an effort to solve this pertinent problem, a three-step approach is proposed, which is based on machine learning techniques to obtain both the position information of users and the trajectory design of UAVs. First, a multi-agent Q-learning-based placement algorithm is proposed for determining the optimal positions of the UAVs based on the initial location of the users. Second, in an effort to determine the mobility information of users based on a real dataset, their position data is collected from Twitter to describe the anonymous user-trajectories in the physical world. In the meantime, an echo state network (ESN) based prediction algorithm is proposed for predicting the future positions of users based on the real dataset. Third, a multi-agent Q-learning-based algorithm is conceived for predicting the position of UAVs in each time slot based on the movement of users. In this algorithm, multiple UAVs act as agents to find optimal actions by interacting with their environment and learn from their mistakes. Additionally, we also prove that the proposed multi-agent Q-learning-based trajectory design and power control algorithm can converge under mild conditions. Numerical results are provided to demonstrate that as the size of the reservoir increases, the proposed ESN approach improves the prediction accuracy. Finally, we demonstrate that the throughput gains of about 17% are achieved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available