论文标题

在SIM2REAL学习上进行权衡:实际学习速度比模拟更快

Trade-off on Sim2Real Learning: Real-world Learning Faster than Simulations

论文作者

Huang, Jingyi, Zhang, Yizheng, Giardina, Fabio, Rosendo, Andre

论文摘要

由于深度神经网络的巨大训练样本需求,深度加固学习(DRL)实验通常在模拟环境中进行。相比之下,基于模型的贝叶斯学习使机器人可以在现实世界中的一些试验中学习良好的政策。尽管迭代次数更少,但贝叶斯方法每次试验的计算成本相对较高,并且这种方法的优势与维度和噪声密切相关。在这里,我们将深层贝叶斯学习算法与无模型的DRL算法进行了比较,同时分析了我们从模拟和现实世界实验中收集的结果。在考虑使用SIM和实际学习的同时,我们的实验表明,即使考虑到计算时间(相对于迭代次数),考虑样品有效的深贝叶斯RL性能也比DRL更好。此外,在模拟中进行的深贝叶斯RL和实验中进行的深贝叶斯RL之间的计算时间差异指出了遍历现实差距的可行途径。我们还表明,SIM与Real之间的混合并不能胜过纯粹的真实方法,这表明现实可以为贝叶斯学习提供最佳的先验知识。机器人每天设计和建造机器人,我们的结果表明,现实世界中的更高学习效率将通过跳过模拟来缩短设计和部署之间的时间。

Deep Reinforcement Learning (DRL) experiments are commonly performed in simulated environments due to the tremendous training sample demands from deep neural networks. In contrast, model-based Bayesian Learning allows a robot to learn good policies within a few trials in the real world. Although it takes fewer iterations, Bayesian methods pay a relatively higher computational cost per trial, and the advantage of such methods is strongly tied to dimensionality and noise. In here, we compare a Deep Bayesian Learning algorithm with a model-free DRL algorithm while analyzing our results collected from both simulations and real-world experiments. While considering Sim and Real learning, our experiments show that the sample-efficient Deep Bayesian RL performance is better than DRL even when computation time (as opposed to number of iterations) is taken in consideration. Additionally, the difference in computation time between Deep Bayesian RL performed in simulation and in experiments point to a viable path to traverse the reality gap. We also show that a mix between Sim and Real does not outperform a purely Real approach, pointing to the possibility that reality can provide the best prior knowledge to a Bayesian Learning. Roboticists design and build robots every day, and our results show that a higher learning efficiency in the real-world will shorten the time between design and deployment by skipping simulations.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源