InferEdge: characterizing edge inference performance

Gabriel Orion Kai Antar, Blesson Varghese*

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Machine learning (ML) inference is a key workload on edge computing resources, facilitated by deployment mechanisms such as Docker containers or WebAssembly. However, the performance characteristics of the deployment mechanisms for edge inference workloads across different target hardware platforms remain unknown. Therefore, this paper introduces InferEdge, a software suite that automates edge inference performance characterization for a range of edge inference models on heterogeneous processors. Experimental results obtained demonstrate that the suitability of each deployment mechanism depends on the workload and target platform, validating the need for a suite that can automate performance characterization and offer insights. InferEdge is available from https://github.com/blessonvar/InferEdge.
Original languageEnglish
Title of host publicationProceedings of the IEEE/ACM international conference on utility and cloud computing (UCC 2025)
Place of PublicationNew York
PublisherACM
Number of pages10
ISBN (Electronic)9798400722851
DOIs
Publication statusPublished - 31 Dec 2025
Event18th IEEE/ACM International Conference on Utility and Cloud Computing - Nantes University, Nantes, France
Duration: 1 Dec 20254 Dec 2025
https://ucc2025.gitlabpages.inria.fr/web/

Conference

Conference18th IEEE/ACM International Conference on Utility and Cloud Computing
Abbreviated titleUCC2025
Country/TerritoryFrance
CityNantes
Period1/12/254/12/25
Internet address

Keywords

  • Edge inference
  • Performance characterization
  • WebAssembly
  • Docker
  • Edge computing

Fingerprint

Dive into the research topics of 'InferEdge: characterizing edge inference performance'. Together they form a unique fingerprint.

Cite this