4.7 Article

Energy-Aware Inference Offloading for DNN-Driven Applications in Mobile Edge Clouds

期刊

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TPDS.2020.3032443

关键词

Artificial intelligence; Cloud computing; 5G mobile communication; Base stations; Task analysis; Mobile handsets; Heuristic algorithms; Inference offloading; mobile edge clouds; approximation and online algorithms

资金

  1. National Natural Science Foundation of China [61802048, 61802047]
  2. fundamental research funds for the central universities in China [DUT19RC(4)035]
  3. DUT-RU Co-Research Center of Advanced ICT for Active Life
  4. Xinghai Scholar Program in Dalian University of Technology, China
  5. Australian Research Council Discovery Project [DP200101985]
  6. National Natural Science Foundation [61972448]

向作者/读者索取更多资源

With the increasing focus on AI applications, DNNs have been successfully used in various areas, requiring significant computational resources. The advancement in 5G and mobile edge computing provides new possibilities for DNN-driven AI applications.
With increasing focus on Artificial Intelligence (AI) applications, Deep Neural Networks (DNNs) have been successfully used in a number of application areas. As the number of layers and neurons in DNNs increases rapidly, significant computational resources are needed to execute a learned DNN model. This ever-increasing resource demand of DNNs is currently met by large-scale data centers with state-of-the-art GPUs. However, increasing availability of mobile edge computing and 5G technologies provide new possibilities for DNN-driven AI applications, especially where these application make use of data sets that are distributed in different locations. One fundamental process of a DNN-driven application in mobile edge clouds is the adoption of inferencing - the process of executing a pre-trained DNN based on newly generated image and video data from mobile devices. We investigate offloading DNN inference requests in a 5G-enabled mobile edge cloud (MEC), with the aim to admit as many inference requests as possible. We propose exact and approximate solutions to the problem of inference offloading in MECs. We also consider dynamic task offloading for inference requests, and devise an online algorithm that can be adapted in real time. The proposed algorithms are evaluated through large-scale simulations and using a real world test-bed implementation. The experimental results demonstrate that the empirical performance of the proposed algorithms outperform their theoretical counterparts and other similar heuristics reported in literature.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据