kth.sePublications
Change search
Link to record
Permanent link

Direct link
Zhang, Jiaojiao
Publications (10 of 10) Show all publications
Zhang, J., He, X., Huang, Y. & Ling, Q. (2025). Byzantine-Robust and Communication-Efficient Personalized Federated Learning. IEEE Transactions on Signal Processing, 73, 26-39
Open this publication in new window or tab >>Byzantine-Robust and Communication-Efficient Personalized Federated Learning
2025 (English)In: IEEE Transactions on Signal Processing, ISSN 1053-587X, E-ISSN 1941-0476, Vol. 73, p. 26-39Article in journal (Refereed) Published
Abstract [en]

This paper explores constrained non-convex personalized federated learning (PFL), in which a group of workers train local models and a global model, under the coordination of a server. To address the challenges of efficient information exchange and robustness against the so-called Byzantine workers, we propose a projected stochastic gradient descent algorithm for PFL that simultaneously ensures Byzantine-robustness and communication efficiency. We implement personalized learning at the workers aided by the global model, and employ a Huber function-based robust aggregation with an adaptive threshold-selecting strategy at the server to reduce the effects of Byzantine attacks. To improve communication efficiency, we incorporate random communication that allows multiple local updates per communication round. We establish the convergence of our algorithm, showing the effects of Byzantine attacks, random communication, and stochastic gradients on the learning error. Numerical experiments demonstrate the superiority of our algorithm in neural network training compared to existing ones.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Servers, Stochastic processes, Signal processing algorithms, Data models, Computational modeling, Vectors, Federated learning, Convergence, Adaptation models, Robustness, Personalized federated learning, communication efficiency, Byzantine-robustness, constrained non-convex optimization
National Category
Signal Processing
Identifiers
urn:nbn:se:kth:diva-358531 (URN)10.1109/TSP.2024.3514802 (DOI)001386428800008 ()2-s2.0-85211976780 (Scopus ID)
Note

QC 20250120

Available from: 2025-01-20 Created: 2025-01-20 Last updated: 2025-01-20Bibliographically approved
Zhang, J., Zhu, L., Fay, D. & Johansson, M. (2025). Locally Differentially Private Online Federated Learning With Correlated Noise. IEEE Transactions on Signal Processing, 73, 1518-1531
Open this publication in new window or tab >>Locally Differentially Private Online Federated Learning With Correlated Noise
2025 (English)In: IEEE Transactions on Signal Processing, ISSN 1053-587X, E-ISSN 1941-0476, Vol. 73, p. 1518-1531Article in journal (Refereed) Published
Abstract [en]

We introduce a locally differentially private (LDP) algorithm for online federated learning that employs temporally correlated noise to improve utility while preserving privacy. To address challenges posed by the correlated noise and local updates with streaming non-IID data, we develop a perturbed iterate analysis that controls the impact of the noise on the utility. Moreover, we demonstrate how the drift errors from local updates can be effectively managed for several classes of nonconvex loss functions. Subject to an (ε, δ)-LDP budget, we establish a dynamic regret bound that quantifies the impact of key parameters and the intensity of changes in the dynamic environment on the learning performance. Numerical experiments confirm the efficacy of the proposed algorithm.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
correlated noise, differential privacy, dynamic regret, Online federated learning
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-363125 (URN)10.1109/TSP.2025.3553355 (DOI)001463431100004 ()2-s2.0-105003029029 (Scopus ID)
Note

QC 20250506

Available from: 2025-05-06 Created: 2025-05-06 Last updated: 2025-05-28Bibliographically approved
Berglund, E., Zhang, J. & Johansson, M. (2025). Soft quasi-Newton: guaranteed positive definiteness by relaxing the secant constraint. Optimization Methods and Software, 1-30
Open this publication in new window or tab >>Soft quasi-Newton: guaranteed positive definiteness by relaxing the secant constraint
2025 (English)In: Optimization Methods and Software, ISSN 1055-6788, E-ISSN 1029-4937, p. 1-30Article in journal (Refereed) Epub ahead of print
Abstract [en]

We propose a novel algorithm, termed soft quasi-Newton (soft QN), for optimization in the presence of bounded noise. Traditional quasi-Newton algorithms are vulnerable to such noise-induced perturbations. To develop a more robust quasi-Newton method, we replace the secant condition in the matrix optimization problem for the Hessian update with a penalty term in its objective and derive a closed-form update formula. A key feature of our approach is its ability to maintain positive definiteness of the Hessian inverse approximation throughout the iterations. Furthermore, we establish the following properties of soft QN: it recovers the BFGS method under specific limits, it treats positive and negative curvature equally, and it is scale invariant. Collectively, these features enhance the efficacy of soft QN in noisy environments. For strongly convex objective functions and Hessian approximations obtained using soft QN, we develop an algorithm that exhibits linear convergence toward a neighborhood of the optimal solution even when gradient and function evaluations are subject to bounded perturbations. Through numerical experiments, we demonstrate that soft QN consistently outperforms state-of-the-art methods across a range of scenarios.

Place, publisher, year, edition, pages
Informa UK Limited, 2025
Keywords
quasi-Newton methods, general bounded noise, secant condition, penalty
National Category
Computational Mathematics
Identifiers
urn:nbn:se:kth:diva-362428 (URN)10.1080/10556788.2025.2475406 (DOI)001449014500001 ()2-s2.0-105000489741 (Scopus ID)
Note

QC 20250425

Available from: 2025-04-15 Created: 2025-04-15 Last updated: 2025-04-25Bibliographically approved
Zhang, J., Hu, J. & Johansson, M. (2024). COMPOSITE FEDERATED LEARNING WITH HETEROGENEOUS DATA. In: 2024 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2024 - Proceedings: . Paper presented at 49th IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2024, Seoul, Korea, Apr 14 2024 - Apr 19 2024 (pp. 8946-8950). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>COMPOSITE FEDERATED LEARNING WITH HETEROGENEOUS DATA
2024 (English)In: 2024 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2024 - Proceedings, Institute of Electrical and Electronics Engineers (IEEE) , 2024, p. 8946-8950Conference paper, Published paper (Refereed)
Abstract [en]

We propose a novel algorithm for solving the composite Federated Learning (FL) problem. This algorithm manages non-smooth regularization by strategically decoupling the proximal operator and communication, and addresses client drift without any assumptions about data similarity. Moreover, each worker uses local updates to reduce the communication frequency with the server and transmits only a d-dimensional vector per communication round. We prove that our algorithm converges linearly to a neighborhood of the optimal solution and demonstrate the superiority of our algorithm over state-of-the-art methods in numerical experiments.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2024
Series
ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings, ISSN 1520-6149
Keywords
Composite federated learning, heterogeneous data, local update
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-348288 (URN)10.1109/ICASSP48485.2024.10447718 (DOI)001396233802047 ()2-s2.0-85195366479 (Scopus ID)
Conference
49th IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2024, Seoul, Korea, Apr 14 2024 - Apr 19 2024
Note

QC 20240626

Part of ISBN 979-835034485-1

Available from: 2024-06-20 Created: 2024-06-20 Last updated: 2025-03-24Bibliographically approved
Zhang, J., Zhu, L. & Johansson, M. (2024). Differentially Private Online Federated Learning with Correlated Noise. In: 2024 IEEE 63rd Conference on Decision and Control, CDC 2024: . Paper presented at 63rd IEEE Conference on Decision and Control, CDC 2024, Milan, Italy, Dec 16 2024 - Dec 19 2024 (pp. 3140-3146). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Differentially Private Online Federated Learning with Correlated Noise
2024 (English)In: 2024 IEEE 63rd Conference on Decision and Control, CDC 2024, Institute of Electrical and Electronics Engineers (IEEE) , 2024, p. 3140-3146Conference paper, Published paper (Refereed)
Abstract [en]

We introduce a novel differentially private algorithm for online federated learning that employs temporally correlated noise to enhance utility while ensuring privacy of continuously released models. To address challenges posed by DP noise and local updates with streaming non-iid data, we develop a perturbed iterate analysis to control the impact of the DP noise on the utility. Moreover, we demonstrate how the drift errors from local updates can be effectively managed under a quasi-strong convexity condition. Subject to an (, δ) DP budget, we establish a dynamic regret bound over the entire time horizon, quantifying the impact of key parameters and the intensity of changes in dynamic environments. Numerical experiments confirm the efficacy of the proposed algorithm.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2024
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-361764 (URN)10.1109/CDC56724.2024.10886177 (DOI)2-s2.0-86000650544 (Scopus ID)
Conference
63rd IEEE Conference on Decision and Control, CDC 2024, Milan, Italy, Dec 16 2024 - Dec 19 2024
Note

Part of ISBN 9798350316339

QC 20250401

Available from: 2025-03-27 Created: 2025-03-27 Last updated: 2025-04-01Bibliographically approved
Zhang, J., Fay, D. & Johansson, M. (2024). DYNAMIC PRIVACY ALLOCATION FOR LOCALLY DIFFERENTIALLY PRIVATE FEDERATED LEARNING WITH COMPOSITE OBJECTIVES. In: 2024 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2024 - Proceedings: . Paper presented at 49th IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2024, Seoul, Korea, Apr 14 2024 - Apr 19 2024 (pp. 9461-9465). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>DYNAMIC PRIVACY ALLOCATION FOR LOCALLY DIFFERENTIALLY PRIVATE FEDERATED LEARNING WITH COMPOSITE OBJECTIVES
2024 (English)In: 2024 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2024 - Proceedings, Institute of Electrical and Electronics Engineers (IEEE) , 2024, p. 9461-9465Conference paper, Published paper (Refereed)
Abstract [en]

This paper proposes a locally differentially private federated learning algorithm for strongly convex but possibly nonsmooth problems that protects the gradients of each worker against an honest but curious server. The proposed algorithm adds artificial noise to the shared information to ensure privacy and dynamically allocates the time-varying noise variance to minimize an upper bound of the optimization error subject to a predefined privacy budget constraint. This allows for an arbitrarily large but finite number of iterations to achieve both privacy protection and utility up to a neighborhood of the optimal solution, removing the need for tuning the number of iterations. Numerical results show the superiority of the proposed algorithm over state-of-the-art methods.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2024
Keywords
dynamic allocation, Federated learning, local differential privacy
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-348291 (URN)10.1109/ICASSP48485.2024.10448141 (DOI)001396233802150 ()2-s2.0-85195409957 (Scopus ID)
Conference
49th IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2024, Seoul, Korea, Apr 14 2024 - Apr 19 2024
Note

QC 20240625 

Part of ISBN [9798350344851]

Available from: 2024-06-20 Created: 2024-06-20 Last updated: 2025-03-26Bibliographically approved
Zhang, J., Hu, J., So, A. M. & Johansson, M. (2024). Nonconvex Federated Learning on Compact Smooth Submanifolds With Heterogeneous Data. In: Advances in Neural Information Processing Systems 37 - 38th Conference on Neural Information Processing Systems, NeurIPS 2024: . Paper presented at 38th Conference on Neural Information Processing Systems, NeurIPS 2024, Vancouver, Canada, Dec 9 2024 - Dec 15 2024. Neural information processing systems foundation, 37
Open this publication in new window or tab >>Nonconvex Federated Learning on Compact Smooth Submanifolds With Heterogeneous Data
2024 (English)In: Advances in Neural Information Processing Systems 37 - 38th Conference on Neural Information Processing Systems, NeurIPS 2024, Neural information processing systems foundation , 2024, Vol. 37Conference paper, Published paper (Refereed)
Abstract [en]

Many machine learning tasks, such as principal component analysis and low-rank matrix completion, give rise to manifold optimization problems. Although there is a large body of work studying the design and analysis of algorithms for manifold optimization in the centralized setting, there are currently very few works addressing the federated setting. In this paper, we consider nonconvex federated learning over a compact smooth submanifold in the setting of heterogeneous client data. We propose an algorithm that leverages stochastic Riemannian gradients and a manifold projection operator to improve computational efficiency, uses local updates to improve communication efficiency, and avoids client drift. Theoretically, we show that our proposed algorithm converges sub-linearly to a neighborhood of a first-order optimal solution by using a novel analysis that jointly exploits the manifold structure and properties of the loss functions. Numerical experiments demonstrate that our algorithm has significantly smaller computational and communication overhead than existing methods.

Place, publisher, year, edition, pages
Neural information processing systems foundation, 2024
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-361952 (URN)2-s2.0-105000497181 (Scopus ID)
Conference
38th Conference on Neural Information Processing Systems, NeurIPS 2024, Vancouver, Canada, Dec 9 2024 - Dec 15 2024
Note

QC 20250409

Available from: 2025-04-03 Created: 2025-04-03 Last updated: 2025-04-09Bibliographically approved
Liu, H., Zhang, J., So, A.-C. M. & Ling, Q. (2023). A Communication-Efficient Decentralized Newton's Method With Provably Faster Convergence. IEEE TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING OVER NETWORKS, 9, 427-441
Open this publication in new window or tab >>A Communication-Efficient Decentralized Newton's Method With Provably Faster Convergence
2023 (English)In: IEEE TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING OVER NETWORKS, ISSN 2373-776X, Vol. 9, p. 427-441Article in journal (Refereed) Published
Abstract [en]

In this article, we consider a strongly convex finite-sum minimization problem over a decentralized network and pro-pose a communication-efficient decentralized Newton's method for solving it. The main challenges in designing such an algorithm come from three aspects: (i) mismatch between local gradients/Hessians and the global ones; (ii) cost of sharing second-order information; (iii) tradeoff among computation and communication. To handle these challenges, we first apply dynamic average consensus (DAC) so that each node is able to use a local gradient approximation and a local Hessian approximation to track the global gradient and Hessian, respectively. Second, since exchanging Hessian approxi-mations is far from communication-efficient, we require the nodes to exchange the compressed ones instead and then apply an error compensation mechanism to correct for the compression noise. Third, we introduce multi-step consensus for exchanging local variables and local gradient approximations to balance between computation and communication. With novel analysis, we establish the globally linear (resp., asymptotically super-linear) convergence rate of the proposed algorithm when m is constant (resp., tends to infinity), where m = 1 is the number of consensus inner steps. To the best of our knowledge, this is the first super-linear conver-gence result for a communication-efficient decentralized Newton's method. Moreover, the rate we establish is provably faster than those of first-order methods. Our numerical results on various applications corroborate the theoretical findings.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2023
Keywords
Decentralized optimization, convergence rate, Newton's method, compressed communication
National Category
Control Engineering
Identifiers
urn:nbn:se:kth:diva-333564 (URN)10.1109/TSIPN.2023.3290397 (DOI)001028957300001 ()2-s2.0-85164436661 (Scopus ID)
Note

QC 20231122

Available from: 2023-08-03 Created: 2023-08-03 Last updated: 2023-11-22Bibliographically approved
He, X., Zhang, J. & Ling, Q. (2023). Byzantine-Robust and Communication-Efficient Personalized Federated Learning. In: ICASSP 2023: 2023 IEEE International Conference on Acoustics, Speech and Signal Processing, Proceedings. Paper presented at 48th IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2023, Rhodes Island, Greece, Jun 4 2023 - Jun 10 2023. Institute of Electrical and Electronics Engineers Inc.
Open this publication in new window or tab >>Byzantine-Robust and Communication-Efficient Personalized Federated Learning
2023 (English)In: ICASSP 2023: 2023 IEEE International Conference on Acoustics, Speech and Signal Processing, Proceedings, Institute of Electrical and Electronics Engineers Inc. , 2023Conference paper, Published paper (Refereed)
Abstract [en]

This paper investigates personalized federated learning, in which a group of workers are coordinated by a server to train correlated local models, in addition to a common global model. This distributed statistical learning problem faces two challenges: efficiency of information exchange between the workers and the server, and robustness to potential malicious messages from the so-called Byzantine workers. We propose a projected stochastic block gradient descent method to address the robustness issue. Therein, each regular worker learns in a personalized manner with the aid of the global model, and the server judiciously aggregates the local models via a Huber function-based descent step. To improve communication efficiency, we allow the regular workers to perform multi-steps of local update per communication round. Convergence of the proposed method is established for non-convex personalized federated learning. Numerical experiments on neural network training validate advantages of the proposed method over the existing ones.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers Inc., 2023
Keywords
Byzantine-robustness, Communication efficiency, Personalized federated learning
National Category
Signal Processing
Identifiers
urn:nbn:se:kth:diva-340373 (URN)10.1109/ICASSP49357.2023.10095468 (DOI)2-s2.0-85177591541 (Scopus ID)
Conference
48th IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2023, Rhodes Island, Greece, Jun 4 2023 - Jun 10 2023
Note

Part of ISBN 9781728163277

QC 20231204

Available from: 2023-12-04 Created: 2023-12-04 Last updated: 2023-12-04Bibliographically approved
Zhang, J., Liu, H., So, A.-C. M. & Ling, Q. (2023). Variance-Reduced Stochastic Quasi-Newton Methods for Decentralized Learning. IEEE Transactions on Signal Processing, 71, 311-326
Open this publication in new window or tab >>Variance-Reduced Stochastic Quasi-Newton Methods for Decentralized Learning
2023 (English)In: IEEE Transactions on Signal Processing, ISSN 1053-587X, E-ISSN 1941-0476, Vol. 71, p. 311-326Article in journal (Refereed) Published
Abstract [en]

In this work, we investigate stochastic quasi-Newton methods for minimizing a finite sum of cost functions over a decentralized network. We first develop a general algorithmic framework, in which each node constructs a local, inexact quasi-Newton direction that asymptotically approaches the global, exact one at each time step. To do so, a local gradient approximation is constructed using dynamic average consensus to track the average of variance-reduced local stochastic gradients over the entire network, followed by a proper local Hessian inverse approximation. We show that under standard convexity and smoothness assumptions on the cost functions, the methods obtained from our framework converge linearly to the optimal solution if the local Hessian inverse approximations used have uniformly bounded positive eigenvalues. To construct the Hessian inverse approximations with the said boundedness property, we design two fully decentralized stochastic quasi-Newton methods-namely, the damped regularized limited-memory DFP (Davidon-Fletcher-Powell) and the damped limited-memory BFGS (Broyden-Fletcher-Goldfarb-Shanno)-which use a fixed moving window of past local gradient approximations and local decision variables to adaptively construct Hessian inverse approximations. A noteworthy feature of these methods is that they do not require extra sampling or communication. Numerical results show that the proposed decentralized stochastic quasi-Newton methods are much faster than the existing decentralized stochastic first-order algorithms.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2023
Keywords
Costs, Convergence, Signal processing algorithms, Approximation algorithms, Cost function, Newton method, Computational complexity, Decentralized optimization, stochastic quasi-Newton methods, variance reduction, damped limited-memory DFP, damped limited-memory BFGS
National Category
Computational Mathematics
Identifiers
urn:nbn:se:kth:diva-325198 (URN)10.1109/TSP.2023.3240652 (DOI)000944455200002 ()2-s2.0-85148468536 (Scopus ID)
Note

QC 20231122

Available from: 2023-04-03 Created: 2023-04-03 Last updated: 2023-11-22Bibliographically approved
Organisations

Search in DiVA

Show all publications