论文标题

在动态环境中进行深度反应性计划

Deep Reactive Planning in Dynamic Environments

论文作者

Ota, Kei, Jha, Devesh K., Onishi, Tadashi, Kanezaki, Asako, Yoshiyasu, Yusuke, Sasaki, Yoko, Mariyama, Toshisada, Nikovski, Daniel

论文摘要

提议的方法的主要新颖性是,它允许机器人学习可以在执行过程中适应环境变化的端到端策略。尽管在RL文献中已经研究了政策的目标调节,但这种方法并不容易扩展到机器人在执行过程中可能会发生变化的情况。这是人类自然能够做的事情。但是,机器人很难学习此类反射(即自然响应动态环境),尤其是当未明确提供目标位置的情况下,而需要通过视觉传感器来感知。在当前的工作中,我们提出了一种可以通过协同方式将传统运动学计划,深度学习和深入的强化学习结合到任意环境中来实现此类行为的方法。我们证明了在模拟中的多个触及和拾取任务以及6多型工业操纵器的实际系统中提出的方法。可以找到描述我们工作的视频\ url {https://youtu.be/he-ew59grpq}。

The main novelty of the proposed approach is that it allows a robot to learn an end-to-end policy which can adapt to changes in the environment during execution. While goal conditioning of policies has been studied in the RL literature, such approaches are not easily extended to cases where the robot's goal can change during execution. This is something that humans are naturally able to do. However, it is difficult for robots to learn such reflexes (i.e., to naturally respond to dynamic environments), especially when the goal location is not explicitly provided to the robot, and instead needs to be perceived through a vision sensor. In the current work, we present a method that can achieve such behavior by combining traditional kinematic planning, deep learning, and deep reinforcement learning in a synergistic fashion to generalize to arbitrary environments. We demonstrate the proposed approach for several reaching and pick-and-place tasks in simulation, as well as on a real system of a 6-DoF industrial manipulator. A video describing our work could be found \url{https://youtu.be/hE-Ew59GRPQ}.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源