论文标题
魔术:在野外学习表达的3D动物
MagicPony: Learning Articulated 3D Animals in the Wild
论文作者
论文摘要
我们考虑了预测3D形状,表达,观点,质地和照明的问题,就像一匹马一样,将单个测试图像作为输入。我们提出了一种称为MagicPony的新方法,该方法纯粹是从对象类别的野外单视图像中学习的,对变形拓扑的假设最少。其核心是铰接形状和外观的隐式表述,结合了神经场和网格的优势。为了帮助该模型理解对象的形状和姿势,我们会通过现成的自我监视视觉变压器捕获的知识,并将其融合到3D模型中。为了在观点估计中克服本地Optima,我们进一步介绍了一种新的观点抽样方案,该方案没有任何额外的培训成本。 Magicpony在这项具有挑战性的任务上的表现优于先前的工作,尽管仅在真实图像上训练了这项挑战,但在重建艺术方面表现出了出色的概括。
We consider the problem of predicting the 3D shape, articulation, viewpoint, texture, and lighting of an articulated animal like a horse given a single test image as input. We present a new method, dubbed MagicPony, that learns this predictor purely from in-the-wild single-view images of the object category, with minimal assumptions about the topology of deformation. At its core is an implicit-explicit representation of articulated shape and appearance, combining the strengths of neural fields and meshes. In order to help the model understand an object's shape and pose, we distil the knowledge captured by an off-the-shelf self-supervised vision transformer and fuse it into the 3D model. To overcome local optima in viewpoint estimation, we further introduce a new viewpoint sampling scheme that comes at no additional training cost. MagicPony outperforms prior work on this challenging task and demonstrates excellent generalisation in reconstructing art, despite the fact that it is only trained on real images.