论文标题
AI授权的IoT设备的协作推断
Collaborative Inference for AI-Empowered IoT Devices
论文作者
论文摘要
传统上,人工智能(AI)技术,尤其是深度学习系统是大规模云服务器的领域,可以访问高计算和能源。尽管如此,在The Internet(IoT)网络中,与现实世界的接口是使用硬件限制并可以通信的边缘设备进行的。为边缘设备收集的数据提供AI处理的常规方法涉及将样本发送到云中,以延迟,通信,连通性和隐私问题为代价。因此,近年来,人们通过利用其沟通能力来建立协作推断,对能够对边缘设备进行AI AID推断的兴趣日益增加。本文回顾了候选策略,以通过协作促进AI向IoT设备的过渡。我们确定需要在不同的移动性和连接性约束中运行的必要性是考虑多个方案的激励因素,这些方案可以大致分为远程推理的方法,即在云上进行推理,以及那些在边缘上推断的方法。我们从推理准确性,通信延迟,隐私和连接性要求方面确定每个策略的关键特征,从而在现有方法之间进行系统的比较。最后,我们提出了由协作推断的概念提出未来的研究挑战和机遇。
Artificial intelligence (AI) technologies, and particularly deep learning systems, are traditionally the domain of large-scale cloud servers, which have access to high computational and energy resources. Nonetheless, in Internet-of-Things (IoT) networks, the interface with the real-world is carried out using edge devices that are limited in hardware and can communicate. The conventional approach to provide AI processing to data collected by edge devices involves sending samples to the cloud, at the cost of latency, communication, connectivity, and privacy concerns. Consequently, recent years have witnessed a growing interest in enabling AI-aided inference on edge devices by leveraging their communication capabilities to establish collaborative inference. This article reviews candidate strategies for facilitating the transition of AI to IoT devices via collaboration. We identify the need to operate in different mobility and connectivity constraints as a motivating factor to consider multiple schemes, which can be roughly divided into methods where inference is done remotely, i.e., on the cloud, and those that infer on the edge. We identify the key characteristics of each strategy in terms of inference accuracy, communication latency, privacy, and connectivity requirements, providing a systematic comparison between existing approaches. We conclude by presenting future research challenges and opportunities arising from the concept of collaborative inference.