Endre søk
RefereraExporteraLink to record
Permanent link

Direct link
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annet format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annet språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
The Case for Hierarchical Deep Learning Inference at the Network Edge
IMDEA Networks Institute, Madrid, Spain.
IMDEA Networks Institute, Madrid, Spain.
IMDEA Networks Institute, Madrid, Spain.
KTH, Skolan för elektroteknik och datavetenskap (EECS), Intelligenta system, Teknisk informationsvetenskap.ORCID-id: 0000-0002-2739-5060
Vise andre og tillknytning
Rekke forfattare: 62023 (engelsk)Inngår i: NetAISys 2023 - Proceedings of the 1st International Workshop on Networked AI Systems, Part of MobiSys 2023, Association for Computing Machinery (ACM) , 2023, s. 13-18Konferansepaper, Publicerat paper (Fagfellevurdert)
Fritextbeskrivning
Abstract [en]

Resource-constrained Edge Devices (EDs), e.g., IoT sensors and microcontroller units, are expected to make intelligent decisions using Deep Learning (DL) inference at the edge of the network. Toward this end, developing tinyML models is an area of active research - DL models with reduced computation and memory storage requirements - that can be embedded on these devices. However, tinyML models have lower inference accuracy. On a different front, DNN partitioning and inference offloading techniques were studied for distributed DL inference between EDs and Edge Servers (ESs). In this paper, we explore Hierarchical Inference (HI), a novel approach proposed in [19] for performing distributed DL inference at the edge. Under HI, for each data sample, an ED first uses a local algorithm (e.g., a tinyML model) for inference. Depending on the application, if the inference provided by the local algorithm is incorrect or further assistance is required from large DL models on edge or cloud, only then the ED offloads the data sample. At the outset, HI seems infeasible as the ED, in general, cannot know if the local inference is sufficient or not. Nevertheless, we present the feasibility of implementing HI for image classification applications. We demonstrate its benefits using quantitative analysis and show that HI provides a better trade-off between offloading cost, throughput, and inference accuracy compared to alternate approaches.

sted, utgiver, år, opplag, sider
Association for Computing Machinery (ACM) , 2023. s. 13-18
Emneord [en]
deep learning, edge computing, hierarchical inference
HSV kategori
Identifikatorer
URN: urn:nbn:se:kth:diva-334523DOI: 10.1145/3597062.3597278ISI: 001119206300003Scopus ID: 2-s2.0-85164295939OAI: oai:DiVA.org:kth-334523DiVA, id: diva2:1790623
Konferanse
1st International Workshop on Networked AI Systems, NetAISys 2023, co-located with ACM MobiSys 2023, Helsinki, Finland, Jun 18 2023
Merknad

Part of ISBN 9798400702129

QC 20230823

Tilgjengelig fra: 2023-08-23 Laget: 2023-08-23 Sist oppdatert: 2025-07-16bibliografisk kontrollert

Open Access i DiVA

Fulltekst mangler i DiVA

Andre lenker

Forlagets fulltekstScopus

Person

Moothedath, Vishnu NarayananGross, James

Søk i DiVA

Av forfatter/redaktør
Behera, Adarsh PrasadMoothedath, Vishnu NarayananGross, James
Av organisasjonen

Søk utenfor DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric

doi
urn-nbn
Totalt: 300 treff
RefereraExporteraLink to record
Permanent link

Direct link
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annet format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annet språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf