EcoFed: efficient communication for DNN partitioning-based federated learning

Di Wu*, Rehmat Ullah, Philip Rodgers, Peter Kilpatrick, Ivor Spence, Blesson Varghese

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

18 Downloads (Pure)

Abstract

Efficiently running federated learning (FL) on resource-constrained devices is challenging since they are required to train computationally intensive deep neural networks (DNN) independently. DNN partitioning-based FL (DPFL) has been proposed as one mechanism to accelerate training where the layers of a DNN (or computation) are offloaded from the device to the server. However, this creates significant communication overheads since the intermediate activation and gradient need to be transferred between the device and the server during training. While current research reduces the communication introduced by DNN partitioning using local loss-based methods, we demonstrate that these methods are ineffective in improving the overall efficiency (communication overhead and training speed) of a DPFL system. This is because they suffer from accuracy degradation and ignore the communication costs incurred when transferring the activation from the device to the server. This article proposes Eco Fed-a communication efficient framework for DPFL systems. Eco Fed-a eliminates the transmission of the gradient by developing pre-trained initialization of the DNN model on the device for the first time. This reduces the accuracy degradation seen in local loss-based methods. In addition, EcoFed proposes a novel replay buffer mechanism and implements a quantization-based compression technique to reduce the transmission of the activation. It is experimentally demonstrated that EcoFed can reduce the communication cost by up to 133× and accelerate training by up to 21× when compared to classic FL. Compared to vanilla DPFL, EcoFed achieves a 16× communication reduction and 2.86× training time speed-up. EcoFed is available from https://github.com/blessonvar/EcoFed .
Original languageEnglish
Article number10380682
Number of pages13
JournalIEEE Transactions on Parallel and Distributed Systems
VolumeEarly Access
DOIs
Publication statusPublished - 4 Jan 2024

Keywords

  • Edge computing
  • Federated learning
  • DNN partitioning
  • Communication efficiency

Fingerprint

Dive into the research topics of 'EcoFed: efficient communication for DNN partitioning-based federated learning'. Together they form a unique fingerprint.

Cite this