4.5 Article

CoEdge: Cooperative DNN Inference With Adaptive Workload Partitioning Over Heterogeneous Edge Devices

Journal

IEEE-ACM TRANSACTIONS ON NETWORKING
Volume 29, Issue 2, Pages 595-608

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TNET.2020.3042320

Keywords

Edge intelligence; cooperative DNN inference; distributed computing; energy efficiency

Funding

  1. National Science Foundation of China [U20A20159, U1711265, 61972432]
  2. Program for Guangdong Introducing Innovative and Entrepreneurial Teams [2017ZT07X355]
  3. Pearl River Talent Recruitment Program [2017GC010465]
  4. U.S. National Science Foundation [IIS-1838024, EEC-1801727, CNS-1950485]

Ask authors/readers for more resources

Recent advances in artificial intelligence have led to the growth of intelligent applications at the network edge. To deploy computationally intensive Deep Neural Networks (DNNs) on resource-constrained edge devices, the cooperative execution mechanism offers a new paradigm for achieving high-performance edge intelligence.
Recent advances in artificial intelligence have driven increasing intelligent applications at the network edge, such as smart home, smart factory, and smart city. To deploy computationally intensive Deep Neural Networks (DNNs) on resource-constrained edge devices, traditional approaches have relied on either offloading workload to the remote cloud or optimizing computation at the end device locally. However, the cloud-assisted approaches suffer from the unreliable and delay-significant wide-area network, and the local computing approaches are limited by the constrained computing capability. Towards high-performance edge intelligence, the cooperative execution mechanism offers a new paradigm, which has attracted growing research interest recently. In this paper, we propose CoEdge, a distributed DNN computing system that orchestrates cooperative DNN inference over heterogeneous edge devices. CoEdge utilizes available computation and communication resources at the edge and dynamically partitions the DNN inference workload adaptive to devices' computing capabilities and network conditions. Experimental evaluations based on a realistic prototype show that CoEdge outperforms status-quo approaches in saving energy with close inference latency, achieving up to 25.5% similar to 66.9% energy reduction for four widely-adopted CNN models.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available