You only look once, but compute twice: service function chaining for low-latency object detection in softwarized networks

10Citations
Citations of this article
29Readers
Mendeley users who have this article in their library.

Abstract

With increasing numbers of computer vision and object detection application scenarios, those requiring ultra-low service latency times have become increasingly prominent; e.g., those for autonomous and connected vehicles or smart city applications. The incorporation of machine learning through the applications of trained models in these scenarios can pose a computational challenge. The softwarization of networks provides opportunities to incorporate computing into the network, increasing flexibility by distributing workloads through offloading from client and edge nodes over in-network nodes to servers. In this article, we present an example for splitting the inference component of the YOLOv2 trained machine learning model between client, network, and service side processing to reduce the overall service latency. Assuming a client has 20% of the server computational resources, we observe a more than 12-fold reduction of service latency when incorporating our service split compared to on-client processing and and an increase in speed of more than 25% compared to performing everything on the server. Our approach is not only applicable to object detection, but can also be applied in a broad variety of machine learning-based applications and services.

Cite

CITATION STYLE

APA

Xiang, Z., Seeling, P., & Fitzek, F. H. P. (2021). You only look once, but compute twice: service function chaining for low-latency object detection in softwarized networks. Applied Sciences (Switzerland), 11(5), 1–14. https://doi.org/10.3390/app11052177

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free