kth.sePublications KTH
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Accelerating Energy-Efficient Federated Learning in Cell-Free Networks With Adaptive Quantization
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Communication Systems, CoS.ORCID iD: 0000-0001-8826-2088
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Information Science and Engineering.ORCID iD: 0000-0002-5407-0835
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Communication Systems, CoS.ORCID iD: 0000-0002-5954-434X
2025 (English)In: IEEE Transactions on Machine Learning in Communications and Networking, ISSN 2831-316X, Vol. 3, p. 761-778Article in journal (Refereed) Published
Abstract [en]

Federated Learning (FL) enables clients to share model parameters instead of raw data, reducing communication overhead. Traditional wireless networks, however, suffer from latency issues when supporting FL. Cell-Free Massive MIMO (CFmMIMO) offers a promising alternative, as it can serve multiple clients simultaneously on shared resources, enhancing spectral efficiency and reducing latency in large-scale FL. Still, communication resource constraints at the client side can impede the completion of FL training. To tackle this issue, we propose a low-latency, energy-efficient FL framework with optimized uplink power allocation for efficient uplink communication. Our approach integrates an adaptive quantization strategy that dynamically adjusts bit allocation for local gradient updates, significantly lowering communication cost. We formulate a joint optimization problem involving FL model updates, local iterations, and power allocation. This problem is solved using sequential quadratic programming (SQP) to balance energy consumption and latency. Moreover, for local model training, clients employ the AdaDelta optimizer, which improves convergence compared to standard SGD, Adam, and RMSProp. We also provide a theoretical analysis of FL convergence under AdaDelta. Numerical results demonstrate that, under equal energy and latency budgets, our power allocation strategy improves test accuracy by up to 7% and 19% compared to Dinkelbach and max-sum rate approaches. Furthermore, across all power allocation methods, our quantization scheme outperforms AQUILA and LAQ, increasing test accuracy by up to 36% and 35%, respectively.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2025. Vol. 3, p. 761-778
Keywords [en]
Quantization (signal), Training, Convergence, Optimization, Adaptation models, Uplink, Resource management, Data models, Costs, Accuracy, Federated learning, cell-free massive MIMO networks, adaptive quantization, power allocation, energy efficiency, straggler effect
National Category
Telecommunications
Identifiers
URN: urn:nbn:se:kth:diva-371455DOI: 10.1109/TMLCN.2025.3583659ISI: 001522923100001OAI: oai:DiVA.org:kth-371455DiVA, id: diva2:2020738
Note

QC 20251211

Available from: 2025-12-11 Created: 2025-12-11 Last updated: 2025-12-11Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full text

Authority records

Mahmoudi, AfsanehXiao, MingBjörnson, Emil

Search in DiVA

By author/editor
Mahmoudi, AfsanehXiao, MingBjörnson, Emil
By organisation
Communication Systems, CoSInformation Science and Engineering
Telecommunications

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 10 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf