论文标题
Procrustean回归网络:从2D注释中学习非刚性对象的3D结构
Procrustean Regression Networks: Learning 3D Structure of Non-Rigid Objects from 2D Annotations
论文作者
论文摘要
我们为训练神经网络提出了一个新颖的框架,该框架能够在仅2D注释作为基础真理中可用时学习非刚性对象的3D信息。最近,有一些方法将非刚性结构(NRSFM)的问题设定结合到了深度学习中,以学习3D结构重建。 NRSFM最重要的难度是同时估算旋转和变形,而先前的作品通过对两者进行回归来处理这一点。在本文中,我们通过提出一个自动确定合适旋转的损失函数来解决这一困难。该网络经过培训的成本函数,这些成本函数包括回归误差和对齐形状的低名项,在培训期间学习了人类骨骼和面部诸如人体骨骼和面部的3D结构,而测试是在单帧的基础上进行的。所提出的方法可以处理缺少条目的输入,实验结果验证了所提出的框架表明,即使基础网络结构非常简单,即使在人类36m,300-vw和超现实数据集的人类36m,300-VW和超现实数据集的最新方法中表现出了出色的重建性能。
We propose a novel framework for training neural networks which is capable of learning 3D information of non-rigid objects when only 2D annotations are available as ground truths. Recently, there have been some approaches that incorporate the problem setting of non-rigid structure-from-motion (NRSfM) into deep learning to learn 3D structure reconstruction. The most important difficulty of NRSfM is to estimate both the rotation and deformation at the same time, and previous works handle this by regressing both of them. In this paper, we resolve this difficulty by proposing a loss function wherein the suitable rotation is automatically determined. Trained with the cost function consisting of the reprojection error and the low-rank term of aligned shapes, the network learns the 3D structures of such objects as human skeletons and faces during the training, whereas the testing is done in a single-frame basis. The proposed method can handle inputs with missing entries and experimental results validate that the proposed framework shows superior reconstruction performance to the state-of-the-art method on the Human 3.6M, 300-VW, and SURREAL datasets, even though the underlying network structure is very simple.