Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Toward efficient resource utilization at edge nodes in federated learning
Blekinge Institute of Technology, Faculty of Computing, Department of Computer Science.ORCID iD: 0000-0002-6309-2892
University of Skövde.
Uppsala University.
Uppsala University.
2024 (English)In: Progress in Artificial Intelligence, ISSN 2192-6352, E-ISSN 2192-6360, Vol. 13, no 2, p. 101-117Article in journal (Refereed) Published
Abstract [en]

Federated learning (FL) enables edge nodes to collaboratively contribute to constructing a global model without sharing their data. This is accomplished by devices computing local, private model updates that are then aggregated by a server. However, computational resource constraints and network communication can become a severe bottleneck for larger model sizes typical for deep learning (DL) applications. Edge nodes tend to have limited hardware resources (RAM, CPU), and the network bandwidth and reliability at the edge is a concern for scaling federated fleet applications. In this paper, we propose and evaluate a FL strategy inspired by transfer learning in order to reduce resource utilization on devices, as well as the load on the server and network in each global training round. For each local model update, we randomly select layers to train, freezing the remaining part of the model. In doing so, we can reduce both server load and communication costs per round by excluding all untrained layer weights from being transferred to the server. The goal of this study is to empirically explore the potential trade-off between resource utilization on devices and global model convergence under the proposed strategy. We implement the approach using the FL framework FEDn. A number of experiments were carried out over different datasets (CIFAR-10, CASA, and IMDB), performing different tasks using different DL model architectures. Our results show that training the model partially can accelerate the training process, efficiently utilizes resources on-device, and reduce the data transmission by around 75% and 53% when we train 25%, and 50% of the model layers, respectively, without harming the resulting global model accuracy. Furthermore, our results demonstrate a negative correlation between the number of participating clients in the training process and the number of layers that need to be trained on each client’s side. As the number of clients increases, there is a decrease in the required number of layers. This observation highlights the potential of the approach, particularly in cross-device use cases. © The Author(s) 2024.

Place, publisher, year, edition, pages
Springer Science+Business Media B.V., 2024. Vol. 13, no 2, p. 101-117
Keywords [en]
Data privacy, Distributed training, Federated learning, Machine learning, Partial training, Training parallelization, Deep learning, Economic and social effects, Learning systems, Edge nodes, Global models, Machine-learning, Model updates, Parallelizations, Resources utilizations
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:bth-26538DOI: 10.1007/s13748-024-00322-3ISI: 001242726300001Scopus ID: 2-s2.0-85195583160OAI: oai:DiVA.org:bth-26538DiVA, id: diva2:1877022
Funder
eSSENCE - An eScience CollaborationAvailable from: 2024-06-25 Created: 2024-06-25 Last updated: 2024-08-05Bibliographically approved

Open Access in DiVA

fulltext(1747 kB)69 downloads
File information
File name FULLTEXT01.pdfFile size 1747 kBChecksum SHA-512
6366ffb02c294172b9dbe4387048b6134f25cbcc1f2d29ba88bc0d0ca3e96448ef27177973d5a70f02787afe5f271638fca751095b44aa4201ba04e6b776d1a8
Type fulltextMimetype application/pdf

Other links

Publisher's full textScopus

Authority records

Alawadi, Sadi

Search in DiVA

By author/editor
Alawadi, Sadi
By organisation
Department of Computer Science
In the same journal
Progress in Artificial Intelligence
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 70 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 493 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf