kth.sePublications KTH
Change search
Link to record
Permanent link

Direct link
Publications (10 of 14) Show all publications
Chen, S., Sun, Y., Qiu, J., Zhang, H. & Chen, H. (2025). Voice-Image Cross-Modal Human Fatigue Detection Based on CNN-ELM Hybrid Model. Journal of Internet Technology, 26(5), 597-606
Open this publication in new window or tab >>Voice-Image Cross-Modal Human Fatigue Detection Based on CNN-ELM Hybrid Model
Show others...
2025 (English)In: Journal of Internet Technology, ISSN 1607-9264, E-ISSN 2079-4029, Vol. 26, no 5, p. 597-606Article in journal (Refereed) Published
Abstract [en]

The deepening of human fatigue will lead to the reduction of exercise ability and work efficiency, the increase of errors and accidents, and even the occurrence of organic diseases. Obviously, it is significant to understand the impact of human fatigue on the health, safe production and safe work of different people. At present, fatigue detection is mostly carried out through EEG and EMG signals. These methods usually have the disadvantages of contact and non-realtime. In response to the aforementioned issues in the process of human fatigue detection, this article effectively applies the visual image analysis method of spectrograms to human fatigue detection and proposes a cross-modal human fatigue detection method based on speech spectral image recognition. First, Mel spectrograms of speech segments in the corpus are extracted, and a fatigue spectrogram data set is established. A deep learning model is established through convolutional neural network (CNN) and extreme learning machine (ELM) for spectral image recognition and fatigue detection. CNN is used to extract features from the input image. The feature mapping will eventually be encoded into a one-dimensional vector and sent to ELM for classification. The experimental results indicate that the speech spectrum image features extracted by this method have better fatigue characterization ability than traditional acoustic features.

Place, publisher, year, edition, pages
Journal of Internet Technology, 2025
Keywords
Fatigue, Mel spectrogram, Cross-modal, CNN, ELM
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-374690 (URN)10.70003/160792642025092605003 (DOI)001586811600008 ()2-s2.0-105026659167 (Scopus ID)
Note

QC 20260116

Available from: 2026-01-08 Created: 2026-01-08 Last updated: 2026-01-16Bibliographically approved
Chen, S., Qiu, J., Zhang, H., Yu, Y., Chen, H. & Sun, Y. (2024). Speech Fatigue Recognition under Small Samples Based on Generative Adversarial Networks and BLSTM. International journal of pattern recognition and artificial intelligence, 38(13), Article ID 2458005.
Open this publication in new window or tab >>Speech Fatigue Recognition under Small Samples Based on Generative Adversarial Networks and BLSTM
Show others...
2024 (English)In: International journal of pattern recognition and artificial intelligence, ISSN 0218-0014, Vol. 38, no 13, article id 2458005Article in journal (Refereed) Published
Abstract [en]

To address the issue of low accuracy in speech fatigue recognition (SFR) under small samples, a method for small-sample SFR based on generative adversarial networks (GANs) is proposed. First, we enable the generator and discriminator to adversarially train and learn the features of the samples, and use the generator to generate high-quality simulated samples to expand our dataset. Then, we transfer discriminator parameters to fatigue identification network to accelerate network training speed. Furthermore, we use a bidirectional long short-term memory network (BLSTM) to further learn temporal fatigue features and improve the recognition rate of fatigue. 720 speech samples from a self-made Chinese speech database (SUSP-SFD) were chosen for training and testing. The results indicate that compared with traditional SFR methods, like convolutional neural networks (CNNs) and long short-term memory network (LSTM), our method improved the SFR rate by about 2.3-6.7%, verifying the effectiveness of the method.

Place, publisher, year, edition, pages
World Scientific Pub Co Pte Ltd, 2024
Keywords
BLSTM, data augmentation, GAN, small samples, speech fatigue recognition, transfer learning
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-366722 (URN)10.1142/S0218001424580059 (DOI)001331698200001 ()2-s2.0-85207634014 (Scopus ID)
Note

QC 20250709

Available from: 2025-07-09 Created: 2025-07-09 Last updated: 2025-07-09Bibliographically approved
Chen, H., Ye, Y., Xiao, M. & Skoglund, M. (2023). Asynchronous Parallel Incremental Block-Coordinate Descent for Decentralized Machine Learning. IEEE Transactions on Big Data, 9(4), 1252-1259
Open this publication in new window or tab >>Asynchronous Parallel Incremental Block-Coordinate Descent for Decentralized Machine Learning
2023 (English)In: IEEE Transactions on Big Data, E-ISSN 2332-7790, Vol. 9, no 4, p. 1252-1259Article in journal (Refereed) Published
Abstract [en]

Machine learning (ML) is a key technique for big-data-driven modelling and analysis of massive Internet of Things (IoT) based intelligent and ubiquitous computing. For fast-increasing applications and data amounts, distributed learning is a promising emerging paradigm since it is often impractical or inefficient to share/aggregate data to a centralized location from distinct ones. This paper studies the problem of training an ML model over decentralized systems, where data are distributed over many user devices and the learning algorithm run on-device, with the aim of relaxing the burden at a central entity/server. Although gossip-based approaches have been used for this purpose in different use cases, they suffer from high communication costs, especially when the number of devices is large. To mitigate this, incremental-based methods are proposed. We first introduce incremental block-coordinate descent (I-BCD) for the decentralized ML, which can reduce communication costs at the expense of running time. To accelerate the convergence speed, an asynchronous parallel incremental BCD (API-BCD) method is proposed, where multiple devices/agents are active in an asynchronous fashion. We derive convergence properties for the proposed methods. Simulation results also show that our API-BCD method outperforms state of the art in terms of running time and communication costs.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2023
Keywords
Decentralized learning, block-coordinate descent, incremental method, asynchronous machine learning
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:kth:diva-334292 (URN)10.1109/TBDATA.2022.3230335 (DOI)001029182700017 ()2-s2.0-85146220732 (Scopus ID)
Note

QC 20230818

Available from: 2023-08-18 Created: 2023-08-18 Last updated: 2023-08-18Bibliographically approved
Chen, H., Huang, S., Zhang, D., Xiao, M., Skoglund, M. & Poor, H. V. (2022). Federated Learning over Wireless IoT Networks with Optimized Communication and Resources. IEEE Internet of Things Journal, 9(17), 16592-16605
Open this publication in new window or tab >>Federated Learning over Wireless IoT Networks with Optimized Communication and Resources
Show others...
2022 (English)In: IEEE Internet of Things Journal, E-ISSN 2327-4662, Vol. 9, no 17, p. 16592-16605Article in journal (Refereed) Published
Abstract [en]

To leverage massive distributed data and computation resources, machine learning in the network edge is considered to be a promising technique especially for large-scale model training. Federated learning (FL), as a paradigm of collaborative learning techniques, has obtained increasing research attention with the benefits of communication efficiency and improved data privacy. Due to the lossy communication channels and limited communication resources (e.g., bandwidth and power), it is of interest to investigate fast responding and accurate FL schemes over wireless systems. Hence, we investigate the problem of jointly optimized communication efficiency and resources for FL over wireless Internet of things (IoT) networks. To reduce complexity, we divide the overall optimization problem into two sub-problems, i.e., the client scheduling problem and the resource allocation problem. To reduce the communication costs for FL in wireless IoT networks, a new client scheduling policy is proposed by reusing stale local model parameters. To maximize successful information exchange over networks, a Lagrange multiplier method is first leveraged by decoupling variables including power variables, bandwidth variables and transmission indicators. Then a linear-search based power and bandwidth allocation method is developed. Given appropriate hyper-parameters, we show that the proposed communication-efficient federated learning (CEFL) framework converges at a strong linear rate. Through extensive experiments, it is revealed that the proposed CEFL framework substantially boosts both the communication efficiency and learning performance of both training loss and test accuracy for FL over wireless IoT networks compared to a basic FL approach with uniform resource allocation.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2022
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:kth:diva-310370 (URN)10.1109/jiot.2022.3151193 (DOI)000846738200091 ()2-s2.0-85124846600 (Scopus ID)
Note

QC 20231108

Available from: 2022-03-29 Created: 2022-03-29 Last updated: 2023-11-08Bibliographically approved
You, Y., You, K., Chen, H. & Oechtering, T. J. (2022). On Data-Driven Self-Calibration for IoT-Based Gas Concentration Monitoring Systems. IEEE Internet of Things Journal, 9(15), 13848-13861
Open this publication in new window or tab >>On Data-Driven Self-Calibration for IoT-Based Gas Concentration Monitoring Systems
2022 (English)In: IEEE Internet of Things Journal, ISSN 2327-4662, Vol. 9, no 15, p. 13848-13861Article in journal (Refereed) Published
Abstract [en]

In this paper, data-driven self-calibration algorithms for the Internet-of-Things-based gas concentration monitoring systems embedded with low-cost gas sensors are designed. The measurement errors are assumed to be caused by imperfect compensation for the variation of sensor component behavior. Specifically, the calibration procedure for the non-dispersive infrared CO2 sensors is developed, for which the temperature dependency is the most dominant drift source. For a single sensor, the hidden Markov model is used to characterize the statistical relationship between different quantities introduced by the physical model that builds on the Beer-Lambert law. For the calibration in the Internet-of-Things-based system, sensors first transmit their belief functions of the true gas concentration level to the cloud. Then the cloud fusion center computes a fused belief function according to certain rules. This belief function is then used as reference for calibrating the sensors. To deal with the case where belief functions highly conflict with each other, a Wasserstein distance based weighted average belief function fusion approach is first proposed as networked calibration algorithm. To achieve more long-term stable calibration results, the networked calibration problem is further formulated as a partially observed Markov decision process problem, and the calibration strategies are derived in a sequential manner. Correspondingly, the deep Q-network approach is applied as a computationally efficient method to solve the proposed Markov decision process problem. The performance of practical designs of the proposed self-calibration algorithms is finally illustrated in numerical experiments utilizing real data.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2022
Keywords
belief function fusion, Calibration, data-driven modeling, deep reinforcement learning., Evidence theory, Gas detectors, Hidden Markov models, Internet of Things, Monitoring, Non-dispersive infrared gas sensor calibration, partially observed Markov decision process, Reinforcement learning, Robot sensing systems
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:kth:diva-310373 (URN)10.1109/jiot.2022.3144934 (DOI)000831217100077 ()2-s2.0-85123691937 (Scopus ID)
Funder
EU, Horizon 2020, 825272
Note

QC 20250611

Available from: 2022-03-29 Created: 2022-03-29 Last updated: 2025-06-11Bibliographically approved
Chen, H. (2022). Reliable and Efficient Distributed Machine Learning. (Doctoral dissertation). Kungliga Tekniska högskolan
Open this publication in new window or tab >>Reliable and Efficient Distributed Machine Learning
2022 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

With the ever-increasing penetration and proliferation of various smart Internet of Things (IoT) applications, machine learning (ML) is envisioned to be a key technique for big-data-driven modelling and analysis. Since massive data generated from these IoT devices are commonly collected and stored in a distributed manner, ML at the networks, e.g., distributed machine learning (DML), has been a promising emerging paradigm, especially for large-scale model training. In this thesis, we explore the optimization and design of DML algorithms under different network conditions. Our main research with regards to DML can be sorted into the following four aspects/papers as detailed below.

In the first part of the thesis, we explore fully-decentralized ML by utilizing alternating direction method of multipliers (ADMM). Specifically, to address the two main critical challenges in DML systems, i.e., communication bottleneck and stragglers (nodes/devices with slow responses), an error-control-coding-based stochastic incremental ADMM (csI-ADMM) is proposed. Given an appropriate mini-batch size, it is proved that the proposed csI-ADMM method has a $O( 1/\sqrt{k})$) convergence rate and $O(1/{\mu ^2})$ communication cost, where $k$ denotes the number of iterations and $\mu$ is the target accuracy.  In addition, tradeoff between the convergence rate and the number of stragglers, as well as the relationship between mini-batch size and number of stragglers, are both theoretically and experimentally analyzed. 

In the second part of the thesis, we investigate the asynchronous approach for fully-decentralized federated learning (FL). Specifically, an asynchronous parallel incremental block-coordinate descent (API-BCD) algorithm is proposed, where multiple nodes/devices are active in an asynchronous fashion to accelerate the convergence speed. The solution convergence of API-BCD is theoretically proved and simulation results demonstrate its superior performance in terms of both running speed and communication costs compared with state-of-the-art algorithms.

The third part of the thesis is devoted to the study of jointly optimizing communication efficiency and wireless resources for FL over wireless networks. Accordingly, an overall optimization problem is formulated, which is divided into two sub-problems, i.e., the client scheduling problem and the resource allocation problem for tractability. More specifically, to reduce the communication costs, a communication-efficient client scheduling policy is proposed by limiting communication exchanges and reusing stale local models. To optimize resource allocation at each communication round of FL training, an optimal solution based on linear search method is derived. The proposed communication-efficient FL (CEFL) algorithm is evaluated both analytically and by simulation. The final part of the thesis is a case study of implementing FL in low Earth orbit (LEO) based satellite communication networks. We investigate four possible architectures of combining ML in LEO-based computing networks. The learning performance of the proposed strategies is evaluated by simulation and results validate that FL-based computing networks can significantly reduce communication overheads and latency.

Abstract [sv]

Med ökat genomslag och spridning av olika smarta Internet of Things (IoT) applikationer, förväntas maskininlärning (ML) bli en nyckelteknik för modellering ochanalys av stora data mängder. Eftersom data från dessa IoT enheter vanligtvis sparaslokalt har ML på nätverksnivå, t.ex. distribuerad maskininlärning (DML), blivit en lovande ny paradigm, särskilt för storskalig modellträning. I denna avhandling utforskarvi optimeringen och designen av DML algoritmer under olika förutsättningar i nätverken. Vår huvudsakliga forskning i hänseende till DML är fördelad i fyra papper,beskrivna enligt nedan.I första delen av denna avhandling tittar vi på fullt decentraliserad ML genom nyttjandet av älternating direction method of multipliers"(ADMM). Mer specifikt föreslårvi en "error-control-coding-baserad stochastic incremental ADMM"(csl-ADMM) föratt tackla de två mest kritiska utmaningarna i DML system, dvs. flaskhalsar i kommunikation och eftersläpare (noder/enheter med långsam respons). Givet en lämpligmini-batch storlek visar vi att den föreslagna csl-ADMM metoden konvergerar medO(1/√k) med en kommunikationskostnad på O(1/µ2), där k är antalet iterationer ochv är sökt noggrannhet. Vi ger även en teoretisk och experimentell analys av sambandetmellan konvergenshastighet och antalet eftersläpare samt sambandet mellan mini-batchstorlek och antalet eftersläpare.I avhandlingens andra del undersöker vi den asynkrona hanteringen av fullt decentraliserad kollaborativ inlärning (FL, eng. Federated Learning). Specifikt föreslår vi enalgoritm för äsynchronous parallel incremental block-coordinate descent"(API-BCD),där flera enheter/noder är asynkront aktiva för att öka konvergens hastigheten. Vi gerteoretiskt bevis för API-BCD lösningens konvergens samt visar simuleringar som demonstrerar dess överlägsna prestanda i termer av både hastighet och kommunikationskostnader jämfört med state-of-the-art algoritmer.Avhandlingens tredje del är en studie i att simultant optimera kommunikations effektivitet och hanteringen av trådlösa resurser för FL över trådlösa nätverk. Ett övergripande optimeringsproblem formuleras, som delas upp i två delproblem, schemaläggning av klienter och ett resursallokerings problem. För att reducera kommunikationskostnaderna, föreslås en effektiv kommunikations policy för schemaläggning av klienter som begränsar kommunikation och återanvändandet av lokala modeller som blivitmindre relevanta med tiden. För att optimera resurs allokeringen i varje kommunikations runda av FL träning, härleds en optimal lösning baserad på en linjär sök metod.Den föreslagna kommunikationseffektiva FL (CEFL, eng. Communication EfficientFL) algoritmen utvärderas både analytiskt och med simulering.Den sista delen av avhandlingen är en fallstudie där FL implementeras i satellitkommunikationsnätverk i låg omloppsbana (LEO, eng. Low Earh Orbit). Vi undersökerfyra möjliga arkitekturer för kombinering av ML i satellitburna datornätverk. Prestandan av de föreslagna strategierna utvärderas via simuleringar och resultaten visar attFL-baserade datornätverk kan anmärkningsvärt minska kommunikationsoverhead ochlatens.

Place, publisher, year, edition, pages
Kungliga Tekniska högskolan, 2022. p. 70
Series
TRITA-EECS-AVL ; 2022:18
Keywords
Distributed machine learning, federated learning, communication efficiency, decentralized optimization
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Research subject
Electrical Engineering
Identifiers
urn:nbn:se:kth:diva-310374 (URN)978-91-8040-179-1 (ISBN)
Public defence
2022-04-28, F3, Lindstedsvägen 26, Stockholm, 13:30 (English)
Opponent
Supervisors
Note

QC 20220404

Available from: 2022-04-04 Created: 2022-04-01 Last updated: 2022-06-25Bibliographically approved
Chen, H., Xiao, M. & Pang, Z. (2022). Satellite-Based Computing Networks with Federated Learning. IEEE wireless communications, 29(1), 78-84
Open this publication in new window or tab >>Satellite-Based Computing Networks with Federated Learning
2022 (English)In: IEEE wireless communications, ISSN 1536-1284, E-ISSN 1558-0687, Vol. 29, no 1, p. 78-84Article in journal, Editorial material (Refereed) Published
Abstract [en]

Driven by the ever increasing penetration and proliferation of data-driven applications, a new generation of wireless communication, the sixth generation (6G) mobile system enhanced by artificial intelligence, has attracted substantial research interests. Among various candidate technologies of 6G, low Earth orbit (LEO) satellites have appealing characteristics of ubiquitous wireless access. However, the costs of satellite communication (SatCom) are still high, relative to their counterparts of ground mobile networks. To support massively interconnected devices with intelligent adaptive learning and reduce expensive traffic in SatCom, we propose federated learning (FL) in LEO-based satellite communication networks. We first review the state-of-the-art LEO-based SatCom and related machine learning (ML) techniques, and then analyze four possible ways of combining ML with satellite networks. The learning performance of the proposed strategies is evaluated by simulation and results reveal that FL-based computing networks improve the performance of communication overheads and latency. Finally, we discuss future research topics along this research direction.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2022
Keywords
federated learning, LEO satellites
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:kth:diva-312272 (URN)10.1109/mwc.008.00353 (DOI)000803106900023 ()2-s2.0-85128546964 (Scopus ID)
Funder
Swedish Foundation for Strategic Research, APR20-0023
Note

QC 20220530

Available from: 2022-05-16 Created: 2022-05-16 Last updated: 2022-06-27Bibliographically approved
Chen, H., Ye, Y., Xiao, M., Skoglund, M. & Poor, H. V. (2021). Coded Stochastic ADMM for Decentralized Consensus Optimization With Edge Computing. IEEE Internet of Things Journal, 8(7), 5360-5373
Open this publication in new window or tab >>Coded Stochastic ADMM for Decentralized Consensus Optimization With Edge Computing
Show others...
2021 (English)In: IEEE Internet of Things Journal, ISSN 2327-4662, Vol. 8, no 7, p. 5360-5373Article in journal (Refereed) Published
Abstract [en]

Big data, including applications with high security requirements, are often collected and stored on multiple heterogeneous devices, such as mobile devices, drones, and vehicles. Due to the limitations of communication costs and security requirements, it is of paramount importance to analyze information in a decentralized manner instead of aggregating data to a fusion center. To train large-scale machine learning models, edge/fog computing is often leveraged as an alternative to centralized learning. We consider the problem of learning model parameters in a multiagent system with data locally processed via distributed edge nodes. A class of minibatch stochastic alternating direction method of multipliers (ADMMs) algorithms is explored to develop the distributed learning model. To address two main critical challenges in distributed learning systems, i.e., communication bottleneck and straggler nodes (nodes with slow responses), error-control-coding-based stochastic incremental ADMM is investigated. Given an appropriate minibatch size, we show that the minibatch stochastic ADMM-based method converges in a rate of O(1/root k), where k denotes the number of iterations. Through numerical experiments, it is revealed that the proposed algorithm is communication efficient, rapidly responding, and robust in the presence of straggler nodes compared with state-of-the-art algorithms.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2021
Keywords
Alternating direction method of multipliers (ADMMs), coded edge computing, consensus optimization, decentralized learning
National Category
Control Engineering
Identifiers
urn:nbn:se:kth:diva-293401 (URN)10.1109/JIOT.2021.3058116 (DOI)000633436600025 ()2-s2.0-85101455750 (Scopus ID)
Note

QC 20210426

Available from: 2021-04-26 Created: 2021-04-26 Last updated: 2022-06-25Bibliographically approved
Araujo, I. M., Natalino, C., Chen, H., De Andrade, M., Cardoso, D. L. & Monti, P. (2020). Availability-Guaranteed Service Function Chain Provisioning with Optional Shared Backups. In: 2020 16TH INTERNATIONAL CONFERENCE ON THE DESIGN OF RELIABLE COMMUNICATION NETWORKS DRCN 2020: . Paper presented at 16th International Conference on the Design of Reliable Communication Networks (DRCN), MAR 24-27, 2020, ELECTR NETWORK. IEEE
Open this publication in new window or tab >>Availability-Guaranteed Service Function Chain Provisioning with Optional Shared Backups
Show others...
2020 (English)In: 2020 16TH INTERNATIONAL CONFERENCE ON THE DESIGN OF RELIABLE COMMUNICATION NETWORKS DRCN 2020, IEEE , 2020Conference paper, Published paper (Refereed)
Abstract [en]

The dynamic provisioning of Service Function Chain (SFC) using Virtual Network Functions (VNFs) is a challenging problem, especially for availability-constrained services. The provisioning of backup resources is often used to ensure that availability requirements are fulfilled. However, the assignment of backup resources should be carefully designed to avoid resource inefficiencies as much as possible. This paper proposes the Optional Backup with Shared Path and Shared Function (OBSPSF) strategy, which aims at improving resource efficiency while fulfilling the availability requirements of SFC requests. The strategy uses optional backup provisioning to ensure that backup resources are assigned only when strictly needed (i.e., when the SFC alone does not meet the availability constraint). Moreover, OBSPSF encourages backup sharing (among both connectivity and backup VNFs) to reduce the backup resource overhead. Results show that the strategy can accommodate orders-of-magnitude more services than benchmark heuristics from the literature.

Place, publisher, year, edition, pages
IEEE, 2020
Series
International Conference on the Design of Reliable Communication Networks DRCN, ISSN 2639-2313
Keywords
Service Function Chaining, Virtualized Network Function, Provisioning, Availability, Shared Protection
National Category
Communication Systems
Identifiers
urn:nbn:se:kth:diva-286188 (URN)10.1109/DRCN48652.2020.1570611128 (DOI)000582415900007 ()2-s2.0-85085503405 (Scopus ID)
Conference
16th International Conference on the Design of Reliable Communication Networks (DRCN), MAR 24-27, 2020, ELECTR NETWORK
Note

QC 20210202

Available from: 2021-02-02 Created: 2021-02-02 Last updated: 2023-03-30Bibliographically approved
Ye, Y., Chen, H., Ma, Z. & Xiao, M. (2020). Decentralized Consensus Optimization Based on Parallel Random Walk. IEEE Communications Letters, 24(2), 391-395
Open this publication in new window or tab >>Decentralized Consensus Optimization Based on Parallel Random Walk
2020 (English)In: IEEE Communications Letters, ISSN 1089-7798, E-ISSN 1558-2558, Vol. 24, no 2, p. 391-395Article in journal (Refereed) Published
Abstract [en]

The alternating direction method of multipliers (ADMM) has recently been recognized as a promising approach for large-scale machine learning models. However, very few results study ADMM from the aspect of communication costs, especially jointly with running time. In this letter, we investigate the communication efficiency and running time of ADMM in solving the consensus optimization problem over decentralized networks. We first review the effort of random walk ADMM (W-ADMM), which reduces communication costs at the expense of running time. To accelerate the convergence speed of W-ADMM, we propose the parallel random walk ADMM (PW-ADMM) algorithm, where multiple random walks are active at the same time. Moreover, to further reduce the running time of PW-ADMM, the intelligent parallel random walk ADMM (IPW-ADMM) algorithm is proposed through integrating the Random Walk with Choice with PW-ADMM. By numerical results from simulations, we demonstrate that the proposed algorithms can be both communication efficient and fast in running speed compared with state-of-the-art methods.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2020
Keywords
Convex functions, Optimization, Convergence, Computational complexity, Clocks, Machine learning, Acceleration, Decentralized network, consensus optimization, alternating direction method of multipliers (ADMM)
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-271713 (URN)10.1109/LCOMM.2019.2955442 (DOI)000519909600034 ()2-s2.0-85079824917 (Scopus ID)
Note

QC 20200421

Available from: 2020-04-21 Created: 2020-04-21 Last updated: 2023-12-11Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0001-7579-822x

Search in DiVA

Show all publications