<div class="csl-bib-body">
<div class="csl-entry">May, D., Tundo, A., Ilager, S., & Brandic, I. (2024). <i>DynaSplit: A Hardware-Software Co-Design Framework for Energy-Aware Inference on Edge</i>. arXiv. https://doi.org/10.48550/arXiv.2410.23881</div>
</div>
-
dc.identifier.uri
http://hdl.handle.net/20.500.12708/226151
-
dc.description.abstract
The deployment of ML models on edge devices is challenged by limited computational resources and energy availability. While split computing enables the decomposition of large neural networks (NNs) and allows partial computation on both edge and cloud devices, identifying the most suitable split layer and hardware configurations is a non-trivial task. This process is in fact hindered by the large configuration space, the non-linear dependencies between software and hardware parameters, the heterogeneous hardware and energy characteristics, and the dynamic workload conditions. To overcome this challenge, we propose DynaSplit, a two-phase framework that dynamically configures parameters across both software (i.e., split layer) and hardware (e.g., accelerator usage, CPU frequency). During the Offline Phase, we solve a multi-objective optimization problem with a meta-heuristic approach to discover optimal settings. During the Online Phase, a scheduling algorithm identifies the most suitable settings for an incoming inference request and configures the system accordingly. We evaluate DynaSplit using popular pre-trained NNs on a real-world testbed. Experimental results show a reduction in energy consumption up to 72% compared to cloud-only computation, while meeting ~90% of user request's latency threshold compared to baselines.
en
dc.language.iso
en
-
dc.subject
Edge AI
en
dc.subject
Split Computing
en
dc.subject
Hardware–Software Co-Design
en
dc.subject
Multi-Objective Optimization
en
dc.title
DynaSplit: A Hardware-Software Co-Design Framework for Energy-Aware Inference on Edge