论文标题
点云场景完成,具有单个RGB-D图像的联合色彩和语义估算
Point Cloud Scene Completion with Joint Color and Semantic Estimation from Single RGB-D Image
论文作者
论文摘要
我们提出了一种深入的增强学习方法,用于在体积指导下完成彩色语义点云场景的渐进式视图介绍,从而从仅带有严重闭塞的单个RGB-D图像实现了高质量的场景重建。我们的方法是端到端的,由三个模块组成:3D场景卷重建,2D RGB-D和分段图像入口以及用于完成的多视图选择。鉴于单个RGB-D图像,我们的方法首先预测其语义分割图,并通过3D卷分支获得体积场景重建,作为下一个视图介绍步骤的指南,该步骤试图弥补丢失的信息;第三步涉及在输入的相同视图下投射卷,使它们串联以完成当前视图RGB-D和分段图,并将所有RGB-D和分段图集成到点云中。由于无法使用被遮挡的区域,因此我们诉诸于A3C网络,以浏览并逐步逐步选择大洞完成的下一个最佳视图,直到现场充分重建,同时保证有效性。共同学习所有步骤,以实现强大而一致的结果。我们通过对3D-Future数据进行广泛的实验进行定性和定量评估,获得了比最先进的结果更好的结果。
We present a deep reinforcement learning method of progressive view inpainting for colored semantic point cloud scene completion under volume guidance, achieving high-quality scene reconstruction from only a single RGB-D image with severe occlusion. Our approach is end-to-end, consisting of three modules: 3D scene volume reconstruction, 2D RGB-D and segmentation image inpainting, and multi-view selection for completion. Given a single RGB-D image, our method first predicts its semantic segmentation map and goes through the 3D volume branch to obtain a volumetric scene reconstruction as a guide to the next view inpainting step, which attempts to make up the missing information; the third step involves projecting the volume under the same view of the input, concatenating them to complete the current view RGB-D and segmentation map, and integrating all RGB-D and segmentation maps into the point cloud. Since the occluded areas are unavailable, we resort to a A3C network to glance around and pick the next best view for large hole completion progressively until a scene is adequately reconstructed while guaranteeing validity. All steps are learned jointly to achieve robust and consistent results. We perform qualitative and quantitative evaluations with extensive experiments on the 3D-FUTURE data, obtaining better results than state-of-the-arts.