kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
A Flexible Framework for Communication-Efficient Machine Learning
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Decision and Control Systems (Automatic Control).ORCID iD: 0000-0003-4473-2011
Stockholm Univ, Dept Comp & Syst Sci, Stockholm, Sweden..
Ericsson, Stockholm, Sweden..
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Decision and Control Systems (Automatic Control).ORCID iD: 0000-0002-2237-2580
2021 (English)In: Thirty-Fifth Aaai Conference On Artificial Intelligence, Thirty-Third Conference On Innovative Applications Of Artificial Intelligence And The Eleventh Symposium On Educational Advances In Artificial Intelligence, Association for the Advancement of Artificial Intelligence , 2021, Vol. 35, p. 8101-8109Conference paper, Published paper (Refereed)
Abstract [en]

With the increasing scale of machine learning tasks, it has become essential to reduce the communication between computing nodes. Early work on gradient compression focused on the bottleneck between CPUs and GPUs, but communication-efficiency is now needed in a variety of different system architectures, from high-performance clusters to energy-constrained IoT devices. In the current practice, compression levels are typically chosen before training and settings that work well for one task may be vastly sub-optimal for another dataset on another architecture. In this paper, we propose a flexible framework which adapts the compression level to the true gradient at each iteration, maximizing the improvement in the objective function that is achieved per communicated bit. Our framework is easy to adapt from one technology to the next by modeling how the communication cost depends on the compression level for the specific technology. Theoretical results and practical experiments indicate that the automatic tuning strategies significantly increase communication efficiency on several state-of-the-art compression schemes.

Place, publisher, year, edition, pages
Association for the Advancement of Artificial Intelligence , 2021. Vol. 35, p. 8101-8109
Series
AAAI Conference on Artificial Intelligence, ISSN 2159-5399
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-301798DOI: 10.1609/aaai.v35i9.16987ISI: 000680423508025Scopus ID: 2-s2.0-85121616102OAI: oai:DiVA.org:kth-301798DiVA, id: diva2:1593849
Conference
35th AAAI Conference on Artificial Intelligence / 33rd Conference on Innovative Applications of Artificial Intelligence / 11th Symposium on Educational Advances in Artificial Intelligence, FEB 02-09, 2021, ELECTR NETWORK
Note

Part of proceedings: ISBN 978-1-57735-866-4, QC 20230117

Available from: 2021-09-14 Created: 2021-09-14 Last updated: 2025-05-27Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Khirirat, SaritJohansson, Mikael

Search in DiVA

By author/editor
Khirirat, SaritJohansson, Mikael
By organisation
Decision and Control Systems (Automatic Control)
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 90 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf