论文标题
基于零拍的3D形状检索的域散布的生成对抗网络
Domain Disentangled Generative Adversarial Network for Zero-Shot Sketch-Based 3D Shape Retrieval
论文作者
论文摘要
基于草图的3D形状检索是一项具有挑战性的任务,这是由于草图和3D形状之间的较大域差异。由于现有方法是在相同类别上进行培训和评估的,因此他们无法有效地识别培训期间未使用的类别。在本文中,我们建议用于基于零绘制的3D检索的新型域分解生成对抗网络(DD-GAN),该域中可以检索训练期间无法访问的不看到类别。具体而言,我们首先通过删除草图和3D形状的学习特征来生成域不变的特征和特定于域特异性特征,在该特征中,使用域,不变的特征可与相应的单词嵌入在一起。然后,我们开发了一个生成的对抗网络,该网络将所见类别的域特异性特征与对齐的域不变特征相结合以合成样品,在这种样品中,使用相应的单词嵌入式生成了不看到类别的合成样本。最后,我们使用看不见类别的综合样本与可见类别的真实样本相结合来训练网络进行检索,以便可以识别出未见类别。为了减少域移位问题,我们利用了未看到的未见样本来增强歧视者的歧视能力。通过鉴别器将生成的样品与未看到的未看到样品区分开来,生成器可以生成更现实的看不见的样本。 SHEREC'13和SHEREC'14数据集的广泛实验表明,我们的方法显着提高了看不见类别的检索性能。
Sketch-based 3D shape retrieval is a challenging task due to the large domain discrepancy between sketches and 3D shapes. Since existing methods are trained and evaluated on the same categories, they cannot effectively recognize the categories that have not been used during training. In this paper, we propose a novel domain disentangled generative adversarial network (DD-GAN) for zero-shot sketch-based 3D retrieval, which can retrieve the unseen categories that are not accessed during training. Specifically, we first generate domain-invariant features and domain-specific features by disentangling the learned features of sketches and 3D shapes, where the domain-invariant features are used to align with the corresponding word embeddings. Then, we develop a generative adversarial network that combines the domain-specific features of the seen categories with the aligned domain-invariant features to synthesize samples, where the synthesized samples of the unseen categories are generated by using the corresponding word embeddings. Finally, we use the synthesized samples of the unseen categories combined with the real samples of the seen categories to train the network for retrieval, so that the unseen categories can be recognized. In order to reduce the domain shift problem, we utilized unlabeled unseen samples to enhance the discrimination ability of the discriminator. With the discriminator distinguishing the generated samples from the unlabeled unseen samples, the generator can generate more realistic unseen samples. Extensive experiments on the SHREC'13 and SHREC'14 datasets show that our method significantly improves the retrieval performance of the unseen categories.