4.7 Article

Toward Decentralized and Collaborative Deep Learning Inference for Intelligent IoT Devices

期刊

IEEE NETWORK
卷 36, 期 1, 页码 59-68

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/MNET.011.2000639

关键词

Collaboration; Robustness; Engines; Resource management; Cloud computing; Real-time systems; Computational modeling

资金

  1. National Key R&D Program of China [2019YFF0301500]
  2. National Natural Science Foundation of China (NSFC) [61671081]
  3. Funds for International Cooperation and Exchange of NSFC [61720106007]
  4. 111 Project [B18008]
  5. Fundamental Research Funds for the Central Universities [2018XKJC01]
  6. BUPT Excellent Ph.D.
  7. Students Foundation [CX2019135]

向作者/读者索取更多资源

This paper proposes a decentralized and collaborative deep learning inference system, DeColla, which migrates DNN computations from the cloud center to IoT devices to improve efficiency and robustness. Experimental results show that DeColla outperforms other methods in terms of latency and resource usage.
Deep learning technologies are empowering IoT devices with an increasing number of intelligent services. However, the contradiction between resource-constrained IoT devices and intensive computing makes it common to transfer data to the cloud center for executing all DNN inference, or dynamically allocate DNN computations between IoT devices and the cloud center. Existing approaches perform a strong dependence on the cloud center, and require the support of a reliable and stable network. Thus, it may directly cause unreliable or even unavailable service in extreme or unstable environments. We propose DeColla, a decentralized and collaborative deep learning inference system for IoT devices, which completely migrates DNN computations from the cloud center to the IoT device side, relying on the collaborative mechanism to accelerate the DNN inference that is difficult for an individual IoT device to accomplish. DeColla uses a parallel acceleration strategy via a DRL-based adaptive allocation for collaborative inference, which aims to improve inference efficiency and robustness. To illustrate the advantages and robustness of DeColla, we built a testbed and employ DeColla to evaluate MobileNet DNN network trained on the ImageNet dataset, and also recognize the object for a mobile web AR application and conduct extensive experiments to analyze the latency, resource usage, and robustness against existing methods. Numerical results show that DeColla outperforms other methods in terms of latency and resource usage, which can especially reduce at least 2.5 times latency than the hierarchical inference method when the collaboration is interrupted abnormally.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据