论文标题

自我缩放的视觉变压器(SEVIT)用于鲁棒的医学图像分类

Self-Ensembling Vision Transformer (SEViT) for Robust Medical Image Classification

论文作者

Almalik, Faris, Yaqub, Mohammad, Nandakumar, Karthik

论文摘要

视觉变压器(VIT)竞争替代卷积神经网络(CNN),以完成医学成像(例如分类和分割)中的各种计算机视觉任务。尽管CNN对对抗攻击的脆弱性是一个众所周知的问题,但最近的作品表明,VIT也容易受到此类攻击的影响,并且在攻击下遭受了重大的绩效降解。 VIT易于精心设计的对抗样品的脆弱性引起了人们对它们在临床环境中的安全性的严重关注。在本文中,我们提出了一种新型的自我缩放方法,以在存在对抗攻击的情况下增强VIT的鲁棒性。拟议的自我振兴视觉变压器(SEVIT)利用了一个事实,即通过VIT的初始块所学到的特征表示形式相对不受对抗性扰动的影响。根据这些中间特征表示,学习多个分类器,并将这些预测与最终VIT分类器的预测相结合可以为对抗性攻击提供鲁棒性。测量各种预测之间的一致性也可以帮助检测对抗样本。对两种方式(胸部X射线和基础镜检查)进行的实验证明了SEVIT体系结构在灰色框中防御各种对抗性攻击的功效(攻击者对目标模型有充分的了解,但没有防御机制)设置。代码:https://github.com/faresmalik/sevit

Vision Transformers (ViT) are competing to replace Convolutional Neural Networks (CNN) for various computer vision tasks in medical imaging such as classification and segmentation. While the vulnerability of CNNs to adversarial attacks is a well-known problem, recent works have shown that ViTs are also susceptible to such attacks and suffer significant performance degradation under attack. The vulnerability of ViTs to carefully engineered adversarial samples raises serious concerns about their safety in clinical settings. In this paper, we propose a novel self-ensembling method to enhance the robustness of ViT in the presence of adversarial attacks. The proposed Self-Ensembling Vision Transformer (SEViT) leverages the fact that feature representations learned by initial blocks of a ViT are relatively unaffected by adversarial perturbations. Learning multiple classifiers based on these intermediate feature representations and combining these predictions with that of the final ViT classifier can provide robustness against adversarial attacks. Measuring the consistency between the various predictions can also help detect adversarial samples. Experiments on two modalities (chest X-ray and fundoscopy) demonstrate the efficacy of SEViT architecture to defend against various adversarial attacks in the gray-box (attacker has full knowledge of the target model, but not the defense mechanism) setting. Code: https://github.com/faresmalik/SEViT

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源