论文标题
单个图像对的结构性动态
Structural-analogy from a Single Image Pair
论文作者
论文摘要
近年来,通过使用深层神经网络,无监督的图像到图像翻译的任务已取得了重大进步。通常,所提出的解决方案了解了两个大型未配对图像集合的特征分布,并能够改变给定图像的外观,同时保持其几何形状完整。在本文中,我们探讨了仅给出一对图像的神经网络理解图像结构的功能。我们试图生成结构上对齐的图像:也就是说,生成一个保持B外观和样式的图像,但具有与A的结构排列相对应。这可以控制产生类比的粒度,这决定了样式和内容之间的概念区别。除结构对齐外,我们的方法还可以用于在其他有条件的生成任务中生成高质量的图像,仅利用图像A和B仅使用图像:引导图像综合,样式和纹理传输,文本翻译以及视频翻译。我们的代码和其他结果可在https://github.com/rmokady/structural-analogy/中获得。
The task of unsupervised image-to-image translation has seen substantial advancements in recent years through the use of deep neural networks. Typically, the proposed solutions learn the characterizing distribution of two large, unpaired collections of images, and are able to alter the appearance of a given image, while keeping its geometry intact. In this paper, we explore the capabilities of neural networks to understand image structure given only a single pair of images, A and B. We seek to generate images that are structurally aligned: that is, to generate an image that keeps the appearance and style of B, but has a structural arrangement that corresponds to A. The key idea is to map between image patches at different scales. This enables controlling the granularity at which analogies are produced, which determines the conceptual distinction between style and content. In addition to structural alignment, our method can be used to generate high quality imagery in other conditional generation tasks utilizing images A and B only: guided image synthesis, style and texture transfer, text translation as well as video translation. Our code and additional results are available in https://github.com/rmokady/structural-analogy/.