Abstract
Machine learning (ML) inference is a key workload on edge computing resources, facilitated by deployment mechanisms such as Docker containers or WebAssembly. However, the performance characteristics of the deployment mechanisms for edge inference workloads across different target hardware platforms remain unknown. Therefore, this paper introduces InferEdge, a software suite that automates edge inference performance characterization for a range of edge inference models on heterogeneous processors. Experimental results obtained demonstrate that the suitability of each deployment mechanism depends on the workload and target platform, validating the need for a suite that can automate performance characterization and offer insights. InferEdge is available from https://github.com/blessonvar/InferEdge.
| Original language | English |
|---|---|
| Title of host publication | Proceedings of the IEEE/ACM international conference on utility and cloud computing (UCC 2025) |
| Place of Publication | New York |
| Publisher | ACM |
| Number of pages | 10 |
| ISBN (Electronic) | 9798400722851 |
| DOIs | |
| Publication status | Published - 31 Dec 2025 |
| Event | 18th IEEE/ACM International Conference on Utility and Cloud Computing - Nantes University, Nantes, France Duration: 1 Dec 2025 → 4 Dec 2025 https://ucc2025.gitlabpages.inria.fr/web/ |
Conference
| Conference | 18th IEEE/ACM International Conference on Utility and Cloud Computing |
|---|---|
| Abbreviated title | UCC2025 |
| Country/Territory | France |
| City | Nantes |
| Period | 1/12/25 → 4/12/25 |
| Internet address |
Keywords
- Edge inference
- Performance characterization
- WebAssembly
- Docker
- Edge computing
Fingerprint
Dive into the research topics of 'InferEdge: characterizing edge inference performance'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver