kth.sePublications
Planned maintenance
A system upgrade is planned for 10/12-2024, at 12:00-13:00. During this time DiVA will be unavailable.
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Zeroth-order algorithms for stochastic distributed nonconvex optimization
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Decision and Control Systems (Automatic Control). (Digital Futures)ORCID iD: 0000-0003-4299-0471
Univ North Texas, Dept Elect Engn, Denton, TX 76203 USA..
Northeastern Univ, State Key Lab Synthet Automat Proc Ind, Shenyang 110819, Peoples R China..
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Decision and Control Systems (Automatic Control). (Digital Futures)ORCID iD: 0000-0001-9940-5929
2022 (English)In: Automatica, ISSN 0005-1098, E-ISSN 1873-2836, Vol. 142, article id 110353Article in journal (Refereed) Published
Abstract [en]

In this paper, we consider a stochastic distributed nonconvex optimization problem with the cost function being distributed over n agents having access only to zeroth-order (ZO) information of the cost. This problem has various machine learning applications. As a solution, we propose two distributed ZO algorithms, in which at each iteration each agent samples the local stochastic ZO oracle at two points with a time-varying smoothing parameter. We show that the proposed algorithms achieve the linear speedup convergence rate O(root p/(nT)) for smooth cost functions under the state-dependent variance assumptions which are more general than the commonly used bounded variance and Lipschitz assumptions, and O(p/(nT)) convergence rate when the global cost function additionally satisfies the Polyak-Lojasiewicz (P-L) condition, where p and T are the dimension of the decision variable and the total number of iterations, respectively. To the best of our knowledge, this is the first linear speedup result for distributed ZO algorithms. It consequently enables systematic processing performance improvements by adding more agents. We also show that the proposed algorithms converge linearly under the relatively bounded second moment assumptions and the P-L condition. We demonstrate through numerical experiments the efficiency of our algorithms on generating adversarial examples from deep neural networks in comparison with baseline and recently proposed centralized and distributed ZO algorithms.

Place, publisher, year, edition, pages
Elsevier BV , 2022. Vol. 142, article id 110353
Keywords [en]
Distributed nonconvex optimization, Gradient-free, Linear speedup, Polyak-Lojasiewicz condition, Stochastic optimization
National Category
Robotics
Identifiers
URN: urn:nbn:se:kth:diva-316308DOI: 10.1016/j.automatica.2022.110353ISI: 000830850700015Scopus ID: 2-s2.0-85129954296OAI: oai:DiVA.org:kth-316308DiVA, id: diva2:1686839
Note

QC 20220811

Available from: 2022-08-11 Created: 2022-08-11 Last updated: 2022-08-11Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Yi, XinleiJohansson, Karl H.

Search in DiVA

By author/editor
Yi, XinleiJohansson, Karl H.
By organisation
Decision and Control Systems (Automatic Control)
In the same journal
Automatica
Robotics

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 58 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf