Ändra sökning
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Reducing Communication Overhead of Federated Learning through Clustering Analysis
Blekinge Tekniska Högskola, Fakulteten för datavetenskaper, Institutionen för datavetenskap.ORCID-id: 0000-0001-6061-0861
Blekinge Tekniska Högskola, Fakulteten för datavetenskaper, Institutionen för datavetenskap.ORCID-id: 0000-0003-3128-191x
Blekinge Tekniska Högskola, Fakulteten för datavetenskaper, Institutionen för datavetenskap.ORCID-id: 0000-0002-3118-5058
2021 (Engelska)Ingår i: 26th IEEE Symposium on Computers and Communications (ISCC 2021), Institute of Electrical and Electronics Engineers (IEEE), 2021Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

Training of machine learning models in a Datacenter, with data originated from edge nodes, incurs high communication overheads and violates a user's privacy. These challenges may be tackled by employing Federated Learning (FL) machine learning technique to train a model across multiple decentralized edge devices (workers) using local data. In this paper, we explore an approach that identifies the most representative updates made by workers and those are only uploaded to the central server for reducing network communication costs. Based on this idea, we propose a FL model that can mitigate communication overheads via clustering analysis of the worker local updates. The Cluster Analysis-based Federated Learning (CA-FL) model is studied and evaluated in human activity recognition (HAR) datasets. Our evaluation results show the robustness of CA-FL in comparison with traditional FL in terms of accuracy and communication costs on both IID and non-IID  cases.

Ort, förlag, år, upplaga, sidor
Institute of Electrical and Electronics Engineers (IEEE), 2021.
Serie
Proceedings of the IEEE Symposium on Computers and Communications, ISSN 15301346
Nyckelord [en]
Federated Learning, Communication Costs, Clustering Analysis, Human Activity Recognition
Nationell ämneskategori
Datavetenskap (datalogi)
Forskningsämne
Datavetenskap
Identifikatorer
URN: urn:nbn:se:bth-22237DOI: 10.1109/ISCC53001.2021.9631391ISI: 000936276000023Scopus ID: 2-s2.0-85123206373ISBN: 9781665427449 (tryckt)OAI: oai:DiVA.org:bth-22237DiVA, id: diva2:1605950
Konferens
26th IEEE Symposium on Computers and Communications, ISCC 2021, Athens, 5 September 2021 through 8 September 2021
Tillgänglig från: 2021-10-26 Skapad: 2021-10-26 Senast uppdaterad: 2025-09-30Bibliografiskt granskad
Ingår i avhandling
1. Resource-Aware and Personalized Federated Learning via Clustering Analysis
Öppna denna publikation i ny flik eller fönster >>Resource-Aware and Personalized Federated Learning via Clustering Analysis
2024 (Engelska)Doktorsavhandling, sammanläggning (Övrigt vetenskapligt)
Abstract [en]

Today’s advancement in Artificial Intelligence (AI) enables training Machine Learning (ML) models on the daily-produced data by connected edge devices. To make the most of the data stored on the device, conventional ML approaches require gathering all individual data sets and transferring them to a central location to train a common model. However, centralizing data incurs significant costs related to communication, network resource utilization, high volume of traffic, and privacy issues. To address the aforementioned challenges, Federated Learning (FL) is employed as a novel approach to train a shared model on decentralized edge devices while preserving privacy. Despite the significant potential of FL, it still requires considerable resources such as time, computational power, energy, and bandwidth availability. More importantly, the computational capabilities of the training devices may vary over time. Furthermore, the devices involved in the training process of FL may have distinct training datasets that differ in terms of their size and distribution. As a result of this, the convergence of the FL models may become unstable and slow. These differences can influence the FL process and ultimately lead to suboptimal model performance within a heterogeneous federated network.

In this thesis, we have tackled several of the aforementioned challenges. Initially, a FL algorithm is proposed that utilizes cluster analysis to address the problem of communication overhead. This issue poses a major bottleneck in FL, particularly for complex models, large-scale applications, and frequent updates. The next research conducted in this thesis involved extending the previous study to include wireless networks (WNs). In WSNs, achieving energy-efficient transmission is a significant challenge due to their limited resources. This has motivated us to continue with a comprehensive overview and classification of the latest advancements in context-aware edge-based AI models, with a specific emphasis on sensor networks. The review has also investigated the associated challenges and motivations for adopting AI techniques, along with an evaluation of current areas of research that need further investigation. To optimize the aggregation of the FL model and alleviate communication expenses, the initial study addressing communication overhead is extended to include a FL-based cluster optimization approach. Furthermore, to reduce the detrimental effect caused by data heterogeneity among edge devices on FL, a new study of group-personalized FL models has been conducted. Finally, taking inspiration from the previously mentioned FL models, techniques for assessing clients' contribution by monitoring and evaluating their behavior during training are proposed. In comparison with the most existing contribution evaluation solutions, the proposed techniques do not require significant computational resources.

The FL algorithms presented in this thesis are assessed on a range of real-world datasets. The extensive experiments demonstrated that the proposed FL techniques are effective and robust. These techniques improve communication efficiency, resource utilization, model convergence speed, and aggregation efficiency, and also reduce data heterogeneity when compared to other state-of-the-art methods.

Ort, förlag, år, upplaga, sidor
Karlskrona: Blekinge Tekniska Högskola, 2024. s. 260
Serie
Blekinge Institute of Technology Doctoral Dissertation Series, ISSN 1653-2090 ; 2024:04
Nyckelord
Federated Learning, Clustering Analysis, Eccentricity Analysis, Non- IID Data, Model Personalization
Nationell ämneskategori
Datavetenskap (datalogi)
Forskningsämne
Datavetenskap
Identifikatorer
urn:nbn:se:bth-26081 (URN)978-91-7295-478-6 (ISBN)
Disputation
2024-05-17, C413A, Karlskrona, 10:00 (Engelska)
Opponent
Handledare
Tillgänglig från: 2024-04-05 Skapad: 2024-04-05 Senast uppdaterad: 2025-09-30Bibliografiskt granskad

Open Access i DiVA

fulltext(1607 kB)1558 nedladdningar
Filinformation
Filnamn FULLTEXT01.pdfFilstorlek 1607 kBChecksumma SHA-512
f5a141d3e511e7d99fc122b53d5b41b25756a9c4de2ef5eabd5937f69b65567937d8a56738fd2f1a1ede2b2f8814beb798aee8c360bb39276d72a06cd02ff6cf
Typ fulltextMimetyp application/pdf

Övriga länkar

Förlagets fulltextScopus

Person

Al-Saedi, Ahmed Abbas MohsinBoeva, VeselkaCasalicchio, Emiliano

Sök vidare i DiVA

Av författaren/redaktören
Al-Saedi, Ahmed Abbas MohsinBoeva, VeselkaCasalicchio, Emiliano
Av organisationen
Institutionen för datavetenskap
Datavetenskap (datalogi)

Sök vidare utanför DiVA

GoogleGoogle Scholar
Totalt: 1564 nedladdningar
Antalet nedladdningar är summan av nedladdningar för alla fulltexter. Det kan inkludera t.ex tidigare versioner som nu inte längre är tillgängliga.

doi
isbn
urn-nbn

Altmetricpoäng

doi
isbn
urn-nbn
Totalt: 973 träffar
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf