kth.sePublications KTH
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
One-Shot Federated Learning with Classifier-Free Diffusion Models
Umeå University, Department of Computing Science, Umeå, Sweden, SE-90187.
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0003-0611-4239
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0003-1114-6040
Umeå University, Department of Computing Science, Umeå, Sweden, SE-90187.
2025 (English)In: 2025 IEEE International Conference on Multimedia and Expo: Journey to the Center of Machine Imagination, ICME 2025 - Conference Proceedings, Institute of Electrical and Electronics Engineers (IEEE) , 2025Conference paper, Published paper (Refereed)
Abstract [en]

Federated learning (FL) enables collaborative learning without data centralization but introduces significant communication costs due to multiple communication rounds between clients and the server. One-shot federated learning (OSFL) addresses this by forming a global model with a single communication round, often relying on the server's model distillation or auxiliary dataset generation - mostly through pre-trained diffusion models (DMs). Existing DM-assisted OSFL methods, however, typically employ classifier-guided DMs, which require training auxiliary classifier models at each client, introducing additional computation overhead. This work introduces OSCAR (One-Shot Federated Learning with Classifier-Free Diffusion Models), a novel OSFL approach that eliminates the need for auxiliary models. OSCAR uses foundation models to devise category-specific data representations at each client which are integrated into a classifier-free diffusion model pipeline for server-side data generation. In our experiments, OSCAR outperforms the state-of-the-art on four benchmark datasets while reducing the communication load by at least 99%.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2025.
Keywords [en]
Diffusion Model, Federated Learning, Foundation Model, One-Shot Federated Learning
National Category
Computer Sciences Computer Systems
Identifiers
URN: urn:nbn:se:kth:diva-373858DOI: 10.1109/ICME59968.2025.11209111Scopus ID: 2-s2.0-105022644542OAI: oai:DiVA.org:kth-373858DiVA, id: diva2:2021620
Conference
2025 IEEE International Conference on Multimedia and Expo, ICME 2025, Nantes, France, June 30 - July 4, 2025
Note

Part of ISBN 9798331594954

QC 20251215

Available from: 2025-12-15 Created: 2025-12-15 Last updated: 2025-12-15Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Jin, ShutongPokorny, Florian T.

Search in DiVA

By author/editor
Jin, ShutongPokorny, Florian T.
By organisation
Robotics, Perception and Learning, RPL
Computer SciencesComputer Systems

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 40 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf