Luger, D., Aral, A., & Brandic, I. (2023). Cost-Aware Neural Network Splitting and Dynamic Rescheduling for Edge Intelligence. ACM Journal on Experimental Algorithmics, 42–47. https://doi.org/10.1145/3578354.3592871
With the rise of IoT devices and the necessity of intelligent applications, inference tasks are often offloaded to the cloud due to the computation limitation of the end devices. Yet, requests to the cloud are costly in terms of latency, and therefore a shift of the computation from the cloud to the network's edge is unavoidable. This shift is called edge intelligence and promises lower latency, among other advantages. However, some algorithms, like deep neural networks, are computationally intensive, even for local edge servers (ES). To keep latency low, such DNNs can be split into two parts and distributed between the ES and the cloud. We present a dynamic scheduling algorithm that takes real-Time parameters like the clock speed of the ES, bandwidth, and latency into account and predicts the optimal splitting point regarding latency. Furthermore, we estimate the overall costs for the ES and cloud during run-Time and integrate them into our prediction and decision models. We present a cost-Aware prediction of the splitting point, which can be tuned with a parameter toward faster response or lower costs.
en
Research facilities:
TRIGA Mark II-Nuklearreaktor Universitäre Service-Einrichtung für Transmissionselektronenmikroskopie Vienna Scientific Cluster Zentrum für Kernspinresonanzspektroskopie Zentrum für Mikro & Nanostrukturen
-
Additional information:
EdgeSys '23: Proceedings of the 6th International Workshop on Edge Systems, Analytics and Networking
-
Research Areas:
Logic and Computation: 50% Computer Engineering and Software-Intensive Systems: 30% Computer Science Foundations: 20%