4.7 Article

Incentive-Driven Deep Reinforcement Learning for Content Caching and D2D Offloading

Journal

IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS
Volume 39, Issue 8, Pages 2445-2460

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/JSAC.2021.3087232

Keywords

Device-to-device communication; Telecommunication network management; Mobile nodes; Cellular networks; Reinforcement learning; Optimization; Data models; D2D offloading; deep reinforcement learning; reverse auction; content caching; real mobility trace

Funding

  1. National Natural Science Foundation of China (NSFC) [61872221]
  2. National Science Foundation (NSF) [CNS 1824440, CNS 1828363, CNS 1757533, CNS 1629746, CNS 1564128]

Ask authors/readers for more resources

An Incentive-driven and Deep Q Network (DQN) based Method, named IDQNM, utilizes a reverse auction as an incentive mechanism to motivate nodes to participate in D2D offloading and content caching in order to maximize the CSP's saving cost.
Offloading cellular traffic via Device-to-Device communication (or D2D offloading) has been proved to be an effective way to ease the traffic burden of cellular networks. However, mobile nodes may not be willing to take part in D2D offloading without proper financial incentives since the data offloading process will incur a lot of resource consumption. Therefore, it is imminent to exploit effective incentive mechanisms to motivate nodes to participate in D2D offloading. Furthermore, the design of the content caching strategy is also crucial to the performance of D2D offloading. In this paper, considering these issues, a novel Incentive-driven and Deep Q Network (DQN) based Method, named IDQNM is proposed, in which the reverse auction is employed as the incentive mechanism. Then, the incentive-driven D2D offloading and content caching process is modeled as Integer Non-Linear Programming (INLP), aiming to maximize the saving cost of the Content Service Provider (CSP). To solve the optimization problem, the content caching method based on a Deep Reinforcement Learning (DRL) algorithm, named DQN is proposed to get the approximate optimal solution, and a standard Vickrey-Clarke-Groves (VCG)-based payment rule is proposed to compensate for mobile nodes' cost. Extensive real trace-driven simulation results demonstrate that the proposed IDQNM greatly outperforms other baseline methods in terms of the CSP's saving cost and the offloading rate in different scenarios.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available