论文标题
机器人可以信任你吗?基于DRL的基于信任驱动的人类引导的方法
Can a Robot Trust You? A DRL-Based Approach to Trust-Driven Human-Guided Navigation
论文作者
论文摘要
已知人类使用各种感知输入来构建其日常环境的认知图。因此,当要求人向特定位置提供方向时,他们的寻路能力将这一认知图转换为方向指示的能力就会受到挑战。由于空间焦虑,口语说明中使用的语言可能是模糊的,而且通常不清楚。为了说明导航指导中的这种不可靠性,我们提出了一种新颖的深入强化学习(DRL)基于信任的机器人导航算法,该算法学习了人类的信任度以执行语言指导的导航任务。我们的方法试图回答一个问题,即机器人是否可以信任人类的航海指导。为此,我们研究了一项培训一项政策,该政策仅使用其自身的机器人信任度量指导,仅使用可信赖的人类指导来驶向目标位置。我们考虑量化基于语言的指示的各种情感特征,并以人类信任度量的形式将它们纳入政策的观察空间。我们将这两种信任指标都用于一个最佳认知推理方案,该方案决定何时何时不信任给定的指导。我们的结果表明,学习策略可以以最佳,时间效率的方式导航环境,而不是执行相同任务的探索方法。我们在模拟和现实环境中展示了结果的功效。
Humans are known to construct cognitive maps of their everyday surroundings using a variety of perceptual inputs. As such, when a human is asked for directions to a particular location, their wayfinding capability in converting this cognitive map into directional instructions is challenged. Owing to spatial anxiety, the language used in the spoken instructions can be vague and often unclear. To account for this unreliability in navigational guidance, we propose a novel Deep Reinforcement Learning (DRL) based trust-driven robot navigation algorithm that learns humans' trustworthiness to perform a language guided navigation task. Our approach seeks to answer the question as to whether a robot can trust a human's navigational guidance or not. To this end, we look at training a policy that learns to navigate towards a goal location using only trustworthy human guidance, driven by its own robot trust metric. We look at quantifying various affective features from language-based instructions and incorporate them into our policy's observation space in the form of a human trust metric. We utilize both these trust metrics into an optimal cognitive reasoning scheme that decides when and when not to trust the given guidance. Our results show that the learned policy can navigate the environment in an optimal, time-efficient manner as opposed to an explorative approach that performs the same task. We showcase the efficacy of our results both in simulation and a real-world environment.