kth.sePublications KTH
Operational message
There are currently operational disruptions. Troubleshooting is in progress.
Change search
Link to record
Permanent link

Direct link
Publications (3 of 3) Show all publications
Weng, S., Xiao, M., Ren, C. & Skoglund, M. (2025). Coded Cooperative Networks for Semi-Decentralized Federated Learning. IEEE Wireless Communications Letters, 14(3), 626-630
Open this publication in new window or tab >>Coded Cooperative Networks for Semi-Decentralized Federated Learning
2025 (English)In: IEEE Wireless Communications Letters, ISSN 2162-2337, E-ISSN 2162-2345, Vol. 14, no 3, p. 626-630Article in journal (Refereed) Published
Abstract [en]

To enhance straggler resilience in federated learning (FL) systems, a semi-decentralized approach has been recently proposed, enabling collaboration between clients. Unlike the existing semi-decentralized schemes, which adaptively adjust the collaboration weight according to the network topology, this letter proposes a deterministic coded network that leverages wireless diversity for semi-decentralized FL without requiring prior information about the entire network. Furthermore, the theoretical analyses of the outage and the convergence rate of the proposed scheme are provided. Finally, the superiority of our proposed method over benchmark methods is demonstrated through comprehensive simulations.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Training, Stochastic processes, Convergence, Wireless networks, Signal to noise ratio, Linear programming, Encoding, Computational modeling, Collaboration, Codes, Semi-decentralized federated learning, wireless network, diversity network code, communication stragglers
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:kth:diva-361615 (URN)10.1109/LWC.2024.3518057 (DOI)001439414200029 ()2-s2.0-86000777837 (Scopus ID)
Note

QC 20250326

Available from: 2025-03-26 Created: 2025-03-26 Last updated: 2025-03-26Bibliographically approved
Weng, S., Ren, C., Xiao, M. & Skoglund, M. (2025). Cooperative Gradient Coding. IEEE Transactions on Communications, 73(12), 13087-13102
Open this publication in new window or tab >>Cooperative Gradient Coding
2025 (English)In: IEEE Transactions on Communications, ISSN 0090-6778, E-ISSN 1558-0857, Vol. 73, no 12, p. 13087-13102Article in journal (Refereed) Published
Abstract [en]

This work studies gradient coding (GC) in the context of distributed training problems with unreliable communication. We propose cooperative GC (CoGC), a novel gradient-sharing-based GC framework that leverages cooperative communication among clients. This approach eliminates the need for dataset replication, making it communication- and computation-efficient and suitable for federated learning (FL). By employing the standard GC decoding mechanism, CoGC yields strictly binary outcomes: the global model is either recovered exactly or the recovery is meaningless, with no intermediate outcomes. This characteristic ensures the optimality of the training and demonstrates strong resilience to client-to-server communication failures. However, due to the limited flexibility of the recovery outcomes, the decoding mechanism may also result in communication inefficiency and hinder convergence, especially when communication channels among clients are in poor condition. To overcome this limitation and further exploit the potential of GC matrices, we propose a complementary decoding mechanism, termed GC<sup>+</sup>, which leverages information that would otherwise be discarded during GC decoding failures. This approach significantly improves system reliability against unreliable communication, as the full recovery<sup>1</sup> of the global model dominates in GC<sup>+</sup>. To conclude, this work establishes solid theoretical frameworks for both CoGC and GC<sup>+</sup>. We assess the system reliability by outage analyses and convergence analyses for each decoding mechanism, along with a rigorous investigation of how outages affect the structure and performance of GC matrices. Finally, the effectiveness of CoGC and GC<sup>+</sup> is validated through extensive simulations.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Complementary decoding mechanism, Convergence, Cooperative gradient coding, Federated learning, Secure Aggregation, Straggler mitigation, Unreliable communication
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-371622 (URN)10.1109/TCOMM.2025.3612589 (DOI)001649704400032 ()2-s2.0-105017454960 (Scopus ID)
Note

QC 20260123

Available from: 2025-10-17 Created: 2025-10-17 Last updated: 2026-01-23Bibliographically approved
Weng, S., Li, C., Xiao, M. & Skoglund, M. (2024). Cooperative Gradient Coding for Semi-Decentralized Federated Learning. In: GLOBECOM 2024 - 2024 IEEE Global Communications Conference: . Paper presented at 2024 IEEE Global Communications Conference, GLOBECOM 2024, Cape Town, South Africa, Dec 8 2024 - Dec 12 2024 (pp. 199-204). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Cooperative Gradient Coding for Semi-Decentralized Federated Learning
2024 (English)In: GLOBECOM 2024 - 2024 IEEE Global Communications Conference, Institute of Electrical and Electronics Engineers (IEEE) , 2024, p. 199-204Conference paper, Published paper (Refereed)
Abstract [en]

Stragglers' effects are known to degrade FL performance. In this paper, we investigate federated learning (FL) over wireless networks in the presence of communication stragglers, where the power-constrained clients collaboratively train a global model by iteratively optimizing a local objective function with their local datasets and transmitting local model updates to the central parameter server (PS) through fading channels. To tackle communication stragglers without dataset sharing or prior information about the network at PS, we propose cooperative gradient coding (CoGC) for semi-decentralized FL to enable the exact global model recovery at PS. Furthermore, we conduct a thorough theoretical analysis of the proposed approach. Namely, an outage analysis of the proposed approach is provided, followed by a convergence analysis based on the failure probability of the global model recovery at PS. Nevertheless, simulation results reveal the superiority of the proposed approach in the presence of stragglers under imbalanced data distribution.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2024
Keywords
communication stragglers, convergence, Federated learning, gradient coding, outages, semi-decentralized network
National Category
Communication Systems
Identifiers
urn:nbn:se:kth:diva-361981 (URN)10.1109/GLOBECOM52923.2024.10901785 (DOI)001511158700034 ()2-s2.0-105000827333 (Scopus ID)
Conference
2024 IEEE Global Communications Conference, GLOBECOM 2024, Cape Town, South Africa, Dec 8 2024 - Dec 12 2024
Note

Part of ISBN 9798350351255

QC 20250409

Available from: 2025-04-03 Created: 2025-04-03 Last updated: 2025-12-08Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0003-0930-7001

Search in DiVA

Show all publications