论文标题
D-UNET:用于图像剪接伪造检测和定位的双编码U-NET
D-Unet: A Dual-encoder U-Net for Image Splicing Forgery Detection and Localization
论文作者
论文摘要
最近,已经提出了许多基于卷积神经网络(CNN)的检测方法用于图像剪接伪造检测。这些检测方法中的大多数都集中在本地补丁或本地对象上。实际上,图像剪接伪造检测是一项全球二元分类任务,可通过图像指纹区分篡改和未受损的区域。但是,一些基于CNN的检测网络几乎没有保留某些特定的图像内容,但是如果包括,则可以提高网络的检测准确性。为了解决这些问题,我们提出了一个新型网络,称为双重编码器U-NET(D-UNET)用于图像剪接伪造检测,该检测采用了未固定的编码器和固定的编码器。未固定的编码器自动了解区分篡改和未受损害区域的图像指纹,而固定的编码器有意提供了有助于网络学习和检测的方向信息。该双重编码器之后是空间金字塔全局提取模块,该模块扩展了D-UNET的全球见解,以更准确地对篡改和未受损的区域进行分类。在对D-UNET和最先进方法的实验比较研究中,D-UNET在图像级和像素级检测中的其他方法优于其他方法,而无需对大量伪造图像进行预训练或训练。此外,这对于不同的攻击是稳定的。
Recently, many detection methods based on convolutional neural networks (CNNs) have been proposed for image splicing forgery detection. Most of these detection methods focus on the local patches or local objects. In fact, image splicing forgery detection is a global binary classification task that distinguishes the tampered and non-tampered regions by image fingerprints. However, some specific image contents are hardly retained by CNN-based detection networks, but if included, would improve the detection accuracy of the networks. To resolve these issues, we propose a novel network called dual-encoder U-Net (D-Unet) for image splicing forgery detection, which employs an unfixed encoder and a fixed encoder. The unfixed encoder autonomously learns the image fingerprints that differentiate between the tampered and non-tampered regions, whereas the fixed encoder intentionally provides the direction information that assists the learning and detection of the network. This dual-encoder is followed by a spatial pyramid global-feature extraction module that expands the global insight of D-Unet for classifying the tampered and non-tampered regions more accurately. In an experimental comparison study of D-Unet and state-of-the-art methods, D-Unet outperformed the other methods in image-level and pixel-level detection, without requiring pre-training or training on a large number of forgery images. Moreover, it was stably robust to different attacks.