kth.sePublications KTH
Change search
Link to record
Permanent link

Direct link
Alternative names
Publications (10 of 426) Show all publications
Huang, X., Wang, C., Tian, X., Li, Z., Zhao, C. & Xiao, M. (2025). A Multi-Hop Semantic Communication System Enhanced by Semantic Importance. IEEE Access, 13, 140685-140693
Open this publication in new window or tab >>A Multi-Hop Semantic Communication System Enhanced by Semantic Importance
Show others...
2025 (English)In: IEEE Access, E-ISSN 2169-3536, Vol. 13, p. 140685-140693Article in journal (Refereed) Published
Abstract [en]

Semantic communication is envisioned as a promising solution for tackling the data transmission challenges in future wireless communication systems. In this paper, we propose a novel semantic importance-enhanced multi-hop semantic communication (SIMSC) framework that integrates semantic communication with a map-and-forward (MF) relaying scheme. To enhance transmission reliability, the framework employs a semantic importance (SI) calculation module to aid in training joint source-channel coding (JSCC) with relays, thereby better protecting the delivery of crucial semantic information over wireless channels with long-distance signal propagation. Different from conventional amplify-and-forward (AF) and decode-and-forward (DF) relaying schemes, the proposed MF scheme leverages a deep neural network (DNN) to adaptively transform the received signal, optimizing it for multi-hop transmission. This enhances robustness and flexibility in varying relay conditions. Numerical results demonstrate that the proposed SIMSC framework performs notably better than conventional solutions that do not utilize semantic coding or relays. Moreover, the MF relaying scheme outperforms the AF and DF schemes, especially when relay distances vary across hops.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Semantic communication, multi-hop communication, semantic importance, image classification
National Category
Telecommunications
Identifiers
urn:nbn:se:kth:diva-373479 (URN)10.1109/ACCESS.2025.3593341 (DOI)001550832200013 ()2-s2.0-105012442705 (Scopus ID)
Note

QC 20251203

Available from: 2025-12-03 Created: 2025-12-03 Last updated: 2025-12-03Bibliographically approved
Li, X., Lin, D., Xiao, Y. & Xiao, M. (2025). A Recursive Puncturing Method for PAC Codes Based on the Partial Order. IEEE Communications Letters, 29(5), 978-982
Open this publication in new window or tab >>A Recursive Puncturing Method for PAC Codes Based on the Partial Order
2025 (English)In: IEEE Communications Letters, ISSN 1089-7798, E-ISSN 1558-2558, Vol. 29, no 5, p. 978-982Article in journal (Refereed) Published
Abstract [en]

The development of polarization-adjusted convolutional (PAC) codes has introduced a class of efficient designs for short packet transmission. In this contribution, aiming at more flexible code length and rate matching for time-varying channel scenarios, a low-complexity puncturing algorithm for PAC codes is proposed. Specifically, we introduce a Gaussian approximation (GA) algorithm for PAC codes and propose a GA-based optimization method for punctured patterns. Building on this, we present a Gaussian inverse mapping method based on partial order, utilizing a recursive approach to construct the initial set, which significantly reduces the search complexity. Subsequently, we develop a recursive puncturing algorithm based on partial order. Finally, we integrate this method with Reed-Muller (RM) rules, further reducing the complexity.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Codes, Picture archiving and communication systems, Polar codes, Gaussian approximation, Convolutional codes, Optimization, Approximation algorithms, Vectors, Indexes, Error probability, Polarization-adjusted convolutional codes, puncturing, partial order
National Category
Telecommunications
Identifiers
urn:nbn:se:kth:diva-364713 (URN)10.1109/LCOMM.2025.3549092 (DOI)001484691800046 ()2-s2.0-86000653645 (Scopus ID)
Note

QC 20250703

Available from: 2025-07-03 Created: 2025-07-03 Last updated: 2025-07-03Bibliographically approved
Mahmoudi, A., Xiao, M. & Björnson, E. (2025). Accelerating Energy-Efficient Federated Learning in Cell-Free Networks With Adaptive Quantization. IEEE Transactions on Machine Learning in Communications and Networking, 3, 761-778
Open this publication in new window or tab >>Accelerating Energy-Efficient Federated Learning in Cell-Free Networks With Adaptive Quantization
2025 (English)In: IEEE Transactions on Machine Learning in Communications and Networking, ISSN 2831-316X, Vol. 3, p. 761-778Article in journal (Refereed) Published
Abstract [en]

Federated Learning (FL) enables clients to share model parameters instead of raw data, reducing communication overhead. Traditional wireless networks, however, suffer from latency issues when supporting FL. Cell-Free Massive MIMO (CFmMIMO) offers a promising alternative, as it can serve multiple clients simultaneously on shared resources, enhancing spectral efficiency and reducing latency in large-scale FL. Still, communication resource constraints at the client side can impede the completion of FL training. To tackle this issue, we propose a low-latency, energy-efficient FL framework with optimized uplink power allocation for efficient uplink communication. Our approach integrates an adaptive quantization strategy that dynamically adjusts bit allocation for local gradient updates, significantly lowering communication cost. We formulate a joint optimization problem involving FL model updates, local iterations, and power allocation. This problem is solved using sequential quadratic programming (SQP) to balance energy consumption and latency. Moreover, for local model training, clients employ the AdaDelta optimizer, which improves convergence compared to standard SGD, Adam, and RMSProp. We also provide a theoretical analysis of FL convergence under AdaDelta. Numerical results demonstrate that, under equal energy and latency budgets, our power allocation strategy improves test accuracy by up to 7% and 19% compared to Dinkelbach and max-sum rate approaches. Furthermore, across all power allocation methods, our quantization scheme outperforms AQUILA and LAQ, increasing test accuracy by up to 36% and 35%, respectively.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Quantization (signal), Training, Convergence, Optimization, Adaptation models, Uplink, Resource management, Data models, Costs, Accuracy, Federated learning, cell-free massive MIMO networks, adaptive quantization, power allocation, energy efficiency, straggler effect
National Category
Telecommunications
Identifiers
urn:nbn:se:kth:diva-371455 (URN)10.1109/TMLCN.2025.3583659 (DOI)001522923100001 ()
Note

QC 20251211

Available from: 2025-12-11 Created: 2025-12-11 Last updated: 2025-12-11Bibliographically approved
Jiang, M., Ye, Z., Xiao, Y., Gao, Y., Xiao, M. & Niyato, D. (2025). ACSNet: A Deep Neural Network for Compound GNSS Jamming Signal Classification. IEEE Transactions on Cognitive Communications and Networking
Open this publication in new window or tab >>ACSNet: A Deep Neural Network for Compound GNSS Jamming Signal Classification
Show others...
2025 (English)In: IEEE Transactions on Cognitive Communications and Networking, E-ISSN 2332-7731Article in journal (Refereed) Epub ahead of print
Abstract [en]

In the global navigation satellite system (GNSS), identifying not only single but also compound jamming signals is crucial for ensuring reliable navigation and positioning, particularly in future wireless communication scenarios such as the space-air-ground integrated network (SAGIN). However, conventional techniques often struggle with low recognition accuracy and high computational complexity, especially under low jamming-to-noise ratio (JNR) conditions. To overcome the challenge of accurately identifying compound jamming signals embedded within GNSS signals, we propose ACSNet, a novel convolutional neural network designed specifically for this purpose. Unlike conventional methods that tend to exhibit lower accuracy and higher computational demands, particularly in low JNR environments, ACSNet addresses these issues by integrating asymmetric convolution blocks, which improve sensitivity to subtle signal variations while reducing the number of parameters by approximately 50% compared to symmetric convolutional designs. Simulations demonstrate that ACSNet significantly improves accuracy in low JNR regions and shows robust resilience to power ratio (PR) variations. It achieves an overall accuracy of 91.84% and a Kappa coefficient (×100) of 90.82, and notably reaches near 100% recognition accuracy when the JNR is greater than or equal to −9 dB, confirming its effectiveness and efficiency for practical GNSS interference management applications.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
compound jamming signal, convolutional neural network, Global navigation satellite system (GNSS), low JNR, PR variation
National Category
Signal Processing
Identifiers
urn:nbn:se:kth:diva-370711 (URN)10.1109/TCCN.2025.3607284 (DOI)2-s2.0-105015891953 (Scopus ID)
Note

QC 20250930

Available from: 2025-09-30 Created: 2025-09-30 Last updated: 2025-09-30Bibliographically approved
Ye, Z., Liao, S., Gao, Y., Fang, S., Xiao, Y., Xiao, M. & Zammit, S. (2025). CA-FedRC: Codebook Adaptation via Federated Reservoir Computing in 5G NR. IEEE Transactions on Vehicular Technology, 74(6), 9995-9999
Open this publication in new window or tab >>CA-FedRC: Codebook Adaptation via Federated Reservoir Computing in 5G NR
Show others...
2025 (English)In: IEEE Transactions on Vehicular Technology, ISSN 0018-9545, E-ISSN 1939-9359, Vol. 74, no 6, p. 9995-9999Article in journal (Refereed) Published
Abstract [en]

With the burgeon deployment of the fifth-generation new radio (5 G NR) networks, the codebook plays a crucial role in enabling the base station (BS) to acquire the channel state information (CSI). Different 5 G NR codebooks incur varying overheads and exhibit performance disparities under diverse channel conditions, necessitating codebook adaptation based on channel conditions to reduce feedback overhead while enhancing performance. However, existing methods of 5 G NR codebooks adaptation require significant overhead for model training and feedback or fall short in performance. To address these limitations, this letter introduces a federated reservoir computing framework designed for efficient codebook adaptation in computationally and feedback resource-constrained mobile devices. This framework utilizes a novel series of indicators as input training data, striking an effective balance between performance and feedback overhead. Compared to conventional models, the proposed codebook adaptation via federated reservoir computing (CA-FedRC), achieves rapid convergence and significant loss reduction in both speed and accuracy. Extensive simulations under various channel conditions demonstrate that our algorithm not only reduces resource consumption of users but also accurately identifies channel types, thereby optimizing the trade-off between spectrum efficiency, computational complexity, and feedback overhead.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
5G mobile communication, Training, Reservoir computing, Precoding, Computational modeling, Signal to noise ratio, Interference, Indexes, Discrete Fourier transforms, Adaptation models, 5G NR, codebook adaptation, federated learning
National Category
Telecommunications
Identifiers
urn:nbn:se:kth:diva-370551 (URN)10.1109/TVT.2025.3542139 (DOI)001513230700017 ()2-s2.0-85217972137 (Scopus ID)
Note

QC 20251007

Available from: 2025-10-07 Created: 2025-10-07 Last updated: 2025-10-07Bibliographically approved
Chen, X., Wang, J., Huang, J., Zeng, M., Zheng, Z. & Xiao, M. (2025). Classification-Oriented Semantic Communication for Internet of Things. In: 2025 IEEE 101st Vehicular Technology Conference, VTC 2025-Spring 2025 - Proceedings: . Paper presented at 101st IEEE Vehicular Technology Conference, VTC 2025-Spring 2025, Oslo, Norway, Jun 17 2025 - Jun 20 2025. Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Classification-Oriented Semantic Communication for Internet of Things
Show others...
2025 (English)In: 2025 IEEE 101st Vehicular Technology Conference, VTC 2025-Spring 2025 - Proceedings, Institute of Electrical and Electronics Engineers (IEEE) , 2025Conference paper, Published paper (Refereed)
Abstract [en]

With the rapid development of the Internet of Things (IoT), the number of connected devices has increased exponentially, bringing significant convenience to various aspects of daily life and business operations. However, communication between IoT devices requires a significant amount of bandwidth, putting a strain on the communication system. To address this challenge, we introduce a classification-oriented semantic communication approach that transmits only essential information. We present a novel end-to-end task-oriented semantic communication model, which efficiently serves the classification task at the receiver. In particular, the proposed model first utilizes a neural network-based semantic encoder to extract classification-related semantic features. A transformer-based semantic decoder is used at the receiver to retrieve semantic features and generate classification results. We further introduce a channel encoder and decoder module to improve the ability of a single model to deal with various channel conditions. Simulation results show that, compared with the traditional method, the proposed scheme achieves higher classification accuracy on the ESC-50 dataset and UrbanSound8K dataset and has better performance for various channel conditions.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
end-to-end training, internet of things, Task-oriented communication
National Category
Communication Systems Computer Sciences
Identifiers
urn:nbn:se:kth:diva-372752 (URN)10.1109/VTC2025-Spring65109.2025.11174625 (DOI)2-s2.0-105019045915 (Scopus ID)
Conference
101st IEEE Vehicular Technology Conference, VTC 2025-Spring 2025, Oslo, Norway, Jun 17 2025 - Jun 20 2025
Note

Part of ISBN 979-8-3315-3147-8

QC 20251113

Available from: 2025-11-13 Created: 2025-11-13 Last updated: 2025-11-13Bibliographically approved
Weng, S., Xiao, M., Ren, C. & Skoglund, M. (2025). Coded Cooperative Networks for Semi-Decentralized Federated Learning. IEEE Wireless Communications Letters, 14(3), 626-630
Open this publication in new window or tab >>Coded Cooperative Networks for Semi-Decentralized Federated Learning
2025 (English)In: IEEE Wireless Communications Letters, ISSN 2162-2337, E-ISSN 2162-2345, Vol. 14, no 3, p. 626-630Article in journal (Refereed) Published
Abstract [en]

To enhance straggler resilience in federated learning (FL) systems, a semi-decentralized approach has been recently proposed, enabling collaboration between clients. Unlike the existing semi-decentralized schemes, which adaptively adjust the collaboration weight according to the network topology, this letter proposes a deterministic coded network that leverages wireless diversity for semi-decentralized FL without requiring prior information about the entire network. Furthermore, the theoretical analyses of the outage and the convergence rate of the proposed scheme are provided. Finally, the superiority of our proposed method over benchmark methods is demonstrated through comprehensive simulations.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Training, Stochastic processes, Convergence, Wireless networks, Signal to noise ratio, Linear programming, Encoding, Computational modeling, Collaboration, Codes, Semi-decentralized federated learning, wireless network, diversity network code, communication stragglers
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:kth:diva-361615 (URN)10.1109/LWC.2024.3518057 (DOI)001439414200029 ()2-s2.0-86000777837 (Scopus ID)
Note

QC 20250326

Available from: 2025-03-26 Created: 2025-03-26 Last updated: 2025-03-26Bibliographically approved
Li, C., Xiao, M. & Skoglund, M. (2025). Coded Robust Aggregation for Distributed Learning under Byzantine Attacks. IEEE Transactions on Information Forensics and Security, 20, 11636-11651
Open this publication in new window or tab >>Coded Robust Aggregation for Distributed Learning under Byzantine Attacks
2025 (English)In: IEEE Transactions on Information Forensics and Security, ISSN 1556-6013, E-ISSN 1556-6021, Vol. 20, p. 11636-11651Article in journal (Refereed) Published
Abstract [en]

In this paper, we investigate the problem of distributed learning (DL) in the presence of Byzantine attacks. For this problem, various robust bounded aggregation (RBA) rules have been proposed at the central server to mitigate the impact of Byzantine attacks. However, current DL methods apply RBA rules for the local gradients from the honest devices and the disruptive information from Byzantine devices, and the learning performance degrades significantly when the local gradients of different devices vary considerably from each other. To overcome this limitation, we propose a new DL method to cope with Byzantine attacks based on coded robust aggregation (CRA-DL). Before training begins, the training data are allocated to the devices redundantly. During training, in each iteration, the honest devices transmit coded gradients to the server computed from the allocated training data, and the server then aggregates the information received from both honest and Byzantine devices using RBA rules. In this way, the global gradient can be approximately recovered at the server to update the global model. Compared with current DL methods applying RBA rules, the improvement of CRA-DL is attributed to the fact that the coded gradients sent by the honest devices are closer to each other. This closeness enhances the robustness of the aggregation against Byzantine attacks, since Byzantine messages tend to be significantly different from those of honest devices in this case. We theoretically analyze the convergence performance of CRA-DL. Finally, we present numerical results to verify the superiority of the proposed method over existing baselines, showing its enhanced learning performance under Byzantine attacks.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Byzantine attacks, convergence analysis, distributed learning, gradient coding, robust aggregation
National Category
Signal Processing
Identifiers
urn:nbn:se:kth:diva-372561 (URN)10.1109/TIFS.2025.3624620 (DOI)001606641400002 ()2-s2.0-105019656740 (Scopus ID)
Note

QC 20251111

Available from: 2025-11-11 Created: 2025-11-11 Last updated: 2025-11-11Bibliographically approved
Li, C., Xiao, M. & Skoglund, M. (2025). Communication-Efficient Semi-Decentralized Federated Learning in the Presence of Stragglers. IEEE Transactions on Communications, 1-1
Open this publication in new window or tab >>Communication-Efficient Semi-Decentralized Federated Learning in the Presence of Stragglers
2025 (English)In: IEEE Transactions on Communications, ISSN 0090-6778, E-ISSN 1558-0857, p. 1-1Article in journal (Refereed) Published
Abstract [en]

In this paper, we consider the problem of federated learning (FL) with devices that have intermittent connectivity to the central server. For this problem, the concept of semi-decentralized FL has been proposed in the literature. This paradigm allows non-straggler devices to relay the gradients computed by the stragglers to the server, and enables realization of gradient coding (GC) to mitigate the negative impact of the stragglers that fail to communicate directly to the central server. However, for GC in semi-decentralized FL, the communication overhead caused by information transmission among the devices is significant. To overcome this shortcoming, inspired by the existing communication-optimal exact consensus algorithm (CECA), we propose a new communication-efficient semi-decentralized FL method (COFFEE). In each round, the devices exchange information by taking a certain number of steps towards communication-optimal exact consensus, ensuring that each device obtains the average of the gradients computed by both its previous neighbors and itself. Afterwards, the non-stragglers transmit the local average result to the server for global aggregation to update the global model. We analyze the convergence performance and the communication overhead of COFFEE analytically. Building on this, to further enhance learning performance under a specific communication overhead, we propose an enhanced version of COFFEE with an adaptive aggregation rule at the central server, referred to as A-COFFEE, which adjusts to the straggler pattern of the devices over training rounds. Experiments are conducted to verify that the proposed methods outperform the baseline methods.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
communication efficiency, federated learning, intermittent connectivity, stragglers
National Category
Control Engineering Communication Systems
Identifiers
urn:nbn:se:kth:diva-370075 (URN)10.1109/TCOMM.2025.3605479 (DOI)2-s2.0-105015207710 (Scopus ID)
Note

QC 20250922

Available from: 2025-09-22 Created: 2025-09-22 Last updated: 2025-09-22Bibliographically approved
Weng, S., Ren, C., Xiao, M. & Skoglund, M. (2025). Cooperative Gradient Coding. IEEE Transactions on Communications
Open this publication in new window or tab >>Cooperative Gradient Coding
2025 (English)In: IEEE Transactions on Communications, ISSN 0090-6778, E-ISSN 1558-0857Article in journal (Refereed) Epub ahead of print
Abstract [en]

This work studies gradient coding (GC) in the context of distributed training problems with unreliable communication. We propose cooperative GC (CoGC), a novel gradient-sharing-based GC framework that leverages cooperative communication among clients. This approach eliminates the need for dataset replication, making it communication- and computation-efficient and suitable for federated learning (FL). By employing the standard GC decoding mechanism, CoGC yields strictly binary outcomes: the global model is either recovered exactly or the recovery is meaningless, with no intermediate outcomes. This characteristic ensures the optimality of the training and demonstrates strong resilience to client-to-server communication failures. However, due to the limited flexibility of the recovery outcomes, the decoding mechanism may also result in communication inefficiency and hinder convergence, especially when communication channels among clients are in poor condition. To overcome this limitation and further exploit the potential of GC matrices, we propose a complementary decoding mechanism, termed GC<sup>+</sup>, which leverages information that would otherwise be discarded during GC decoding failures. This approach significantly improves system reliability against unreliable communication, as the full recovery<sup>1</sup> of the global model dominates in GC<sup>+</sup>. To conclude, this work establishes solid theoretical frameworks for both CoGC and GC<sup>+</sup>. We assess the system reliability by outage analyses and convergence analyses for each decoding mechanism, along with a rigorous investigation of how outages affect the structure and performance of GC matrices. Finally, the effectiveness of CoGC and GC<sup>+</sup> is validated through extensive simulations.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Complementary decoding mechanism, Convergence, Cooperative gradient coding, Federated learning, Secure Aggregation, Straggler mitigation, Unreliable communication
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-371622 (URN)10.1109/TCOMM.2025.3612589 (DOI)2-s2.0-105017454960 (Scopus ID)
Note

QC 20251017

Available from: 2025-10-17 Created: 2025-10-17 Last updated: 2025-10-17Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0002-5407-0835

Search in DiVA

Show all publications