4.6 Article

Privacy-Aware Resource Sharing in Cross-Device Federated Model Training for Collaborative Predictive Maintenance

Journal

IEEE ACCESS
Volume 9, Issue -, Pages 120367-120379

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/ACCESS.2021.3108839

Keywords

Collaboration; Predictive models; Manufacturing; Data models; Training; Servers; Image edge detection; SplitNN; federated learning; predictive maintenance; Industry 40

Ask authors/readers for more resources

Industry 4.0 has transformed modern industrial assets into a valuable source of data for optimizing operations, ensuring efficiency, and reducing maintenance costs. Federated Learning offers a mechanism to mine distributed data in a privacy-preserving manner, providing new opportunities for collaborative asset management. However, challenges such as device heterogeneity and scalability need to be addressed to fully integrate FL into predictive maintenance applications.
The proliferation of Industry 4.0 has made modern industrial assets a rich source of data that can be leveraged to optimise operations, ensure efficiency, and minimise maintenance costs. The availability of data is advantageous for asset management, however, attempts to maximise the value of this data often fall short due to additional constraints, such as privacy concerns and data stored in distributed silos that is difficult to access and share. Federated Learning (FL) has been explored to address these challenges and has been demonstrated to provide a mechanism that allows highly distributed data to be mined in a privacy-preserving manner and offering new opportunities for a collaborative approach to asset management. Despite the benefits, FL has some challenges that need to be overcome to make it fully compatible for asset management or more specifically predictive maintenance applications. FL requires a set of clients that participate in the model training process, however, orchestration, device heterogeneity and scalability can hinder the speed and accuracy in the context of collaborative predictive maintenance. To address this challenge, this work proposes a split-learning-based framework (SplitPred) that enables FL clients to maximise available resources within their local network without compromising the benefits of a FL approach (i.e., privacy and shared learning). Experiments performed on the benchmark C-MAPSS data-set demonstrate the advantage of applying SplitPred in the FL process in terms of efficient use of resources, i.e., model convergence time, accuracy, and network load.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available