kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
A Computation and Energy Efficient Hardware Architecture for SSL Acceleration
Fudan Univ., Shanghai, China.
Univ. of Pittsburgh, Pittsburgh, USA.
Fudan Univ., Shanghai, China.
Guangdong Institute of Intelligence Science and Technology, Zhuhai, Guangdong, China, Guangdong.
Show others and affiliations
2025 (English)In: ASP-DAC 2025 - 30th Asia and South Pacific Design Automation Conference, Proceedings, Association for Computing Machinery (ACM) , 2025, p. 23-29Conference paper, Published paper (Refereed)
Abstract [en]

In Computer Vision (CV), the deployment of Convolutional Neural Networks (CNNs) is often hindered by their substantial computational requirements and large labeled datasets. Self-supervised learning (SSL) serves as an effective approach to reducing the reliance on labeled data with the option of augmentation methods to infer and train CNNs. Excluding irrelevant features accelerates learning and improves optimization. We propose a Field-Programmable Gate Array (FPGA)-based hardware accelerator architecture tailored for SSL framework, leveraging its parallelism and reconfigurability to expedite block matching, optimize sparse convolutions, and manage data reuse, significantly improving resource and energy efficiency. The implementation and evaluation of our work on Xilinx ZCU102 FPGA working at 200 MHz confirm that the similarity finding part's FPGA accelerations with a low hardware overhead generates a latency of 0.0106 seconds, surpassing GPU and CPU, and in the sparse CNN's FPGA acceleration part, with the processing of VGG16 and ResNet50, compared with the related FPGA-based works, our design claims a maximum of 3.08× throughput improvement and 1.5× in energy efficiency.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM) , 2025. p. 23-29
Keywords [en]
CNN, computation efficiency, energy efficiency, FPGA acceleration, self-supervised learning
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-361963DOI: 10.1145/3658617.3697548Scopus ID: 2-s2.0-105000311281OAI: oai:DiVA.org:kth-361963DiVA, id: diva2:1949636
Conference
30th Asia and South Pacific Design Automation Conference, ASP-DAC 2025, Tokyo, Japan, January 20-23, 2025
Note

Part of ISBN 9798400706356

QC 20250404

Available from: 2025-04-03 Created: 2025-04-03 Last updated: 2025-04-04Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Xu, Jiawei

Search in DiVA

By author/editor
Xu, Jiawei
By organisation
Electronics and Embedded systems
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 25 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf