论文标题
具有成本效益的机器学习推理卸载用于边缘计算
Cost-effective Machine Learning Inference Offload for Edge Computing
论文作者
论文摘要
由于生成大量数据,因此边缘的计算越来越重要。这在将所有数据运输到远程数据中心和云方面构成了挑战,可以在其中处理和分析它们。另一方面,利用边缘数据对于提供数据驱动和基于机器学习的应用程序至关重要,如果挑战(例如设备功能,连接性和异质性)可以得到缓解。机器学习应用程序非常密集,需要处理大量数据。但是,在计算资源,电源,存储和网络连接方面,边缘设备通常受到资源约束。因此,限制了他们有效,准确地运行最先进的深神经网络(DNN)模型的潜力,这些模型变得越来越大,更复杂。本文提出了一种新颖的卸载机制,它利用安装的基本本地(EDGE)计算资源。所提出的机制使边缘设备可以卸载重型和计算密集型工作负载到边缘节点,而不是使用远程云。我们的卸载机制已被原型和测试,并使用最先进的人和对象检测DNN模型,用于移动机器人和视频监视应用程序。与基于云的卸载策略相比,在准确性和延迟方面,该性能显示出显着的增长。
Computing at the edge is increasingly important since a massive amount of data is generated. This poses challenges in transporting all that data to the remote data centers and cloud, where they can be processed and analyzed. On the other hand, harnessing the edge data is essential for offering data-driven and machine learning-based applications, if the challenges, such as device capabilities, connectivity, and heterogeneity can be mitigated. Machine learning applications are very compute-intensive and require processing of large amount of data. However, edge devices are often resources-constrained, in terms of compute resources, power, storage, and network connectivity. Hence, limiting their potential to run efficiently and accurately state-of-the art deep neural network (DNN) models, which are becoming larger and more complex. This paper proposes a novel offloading mechanism by leveraging installed-base on-premises (edge) computational resources. The proposed mechanism allows the edge devices to offload heavy and compute-intensive workloads to edge nodes instead of using remote cloud. Our offloading mechanism has been prototyped and tested with state-of-the art person and object detection DNN models for mobile robots and video surveillance applications. The performance shows a significant gain compared to cloud-based offloading strategies in terms of accuracy and latency.