论文标题
深度非视线重建
Deep Non-Line-of-Sight Reconstruction
论文作者
论文摘要
近年来,人们对直接视线的成像进行成像的兴趣激增。最突出的技术依赖于时间分辨的光学脉冲响应,该响应是通过用超短光脉冲照亮弥漫性壁,并使用超快时间分辨成像仪观察多弹力间接反射。但是,从这些数据中重建几何形状是一个复杂的非线性反问题,它带来了实质性的计算需求。在本文中,我们采用卷积进料网络来有效地解决重建问题,同时保持良好的重建质量。具体而言,我们设计了一个量身定制的自动编码器体系结构,训练有素的端到端,将瞬态图像直接映射到深度图表示。使用有效的瞬态渲染器进行训练,用于弥漫三弹性间接光传输,从而使网络迅速生成大量的培训数据。我们研究了我们的方法对各种合成和实验数据集的性能及其对培训数据和增强策略的选择以及建筑特征的依赖。我们证明了我们的馈电网络,即使仅根据合成数据进行了训练,也可以从SPAD传感器中测量数据,并能够获得与基于模型的重建方法竞争的结果。
The recent years have seen a surge of interest in methods for imaging beyond the direct line of sight. The most prominent techniques rely on time-resolved optical impulse responses, obtained by illuminating a diffuse wall with an ultrashort light pulse and observing multi-bounce indirect reflections with an ultrafast time-resolved imager. Reconstruction of geometry from such data, however, is a complex non-linear inverse problem that comes with substantial computational demands. In this paper, we employ convolutional feed-forward networks for solving the reconstruction problem efficiently while maintaining good reconstruction quality. Specifically, we devise a tailored autoencoder architecture, trained end-to-end, that maps transient images directly to a depth map representation. Training is done using an efficient transient renderer for diffuse three-bounce indirect light transport that enables the quick generation of large amounts of training data for the network. We examine the performance of our method on a variety of synthetic and experimental datasets and its dependency on the choice of training data and augmentation strategies, as well as architectural features. We demonstrate that our feed-forward network, even though it is trained solely on synthetic data, generalizes to measured data from SPAD sensors and is able to obtain results that are competitive with model-based reconstruction methods.