论文标题

“该制造商的名称”。在训练深度学习模型时,将图像采集偏见与任务复杂性有关:头部CT实验

"Name that manufacturer". Relating image acquisition bias with task complexity when training deep learning models: experiments on head CT

论文作者

Biondetti, Giorgio Pietro, Gauriau, Romane, Bridge, Christopher P., Lu, Charles, Andriole, Katherine P.

论文摘要

随着对医学图像应用机器学习技术的兴趣继续以快速的速度增长,因此开始开发和部署用于临床应用的模型。在临床AI模型开发生命周期(由Lu等人[1]描述)中,机器学习科学家和临床医生的关键阶段是数据同类的适当设计和收集。在此步骤中,识别各种形式的偏见和分布变化的能力至关重要。尽管很难考虑所有潜在的偏见来源,但可以开发技术来识别特定类型的偏见以减轻其影响。在这项工作中,我们分析了扫描仪制造商在数据集中的分布如何有助于深度学习模型的整体偏见。我们评估了分类和分割任务的卷积神经网络(CNN),特别是两个最先进的模型:用于分类的Resnet [2]和用于分割的U-NET [3]。我们证明,CNN可以学会区分成像扫描仪制造商,并且这种偏见可以显着影响分类和分割任务的模型性能。通过创建一个模仿或多或少微妙病变存在的大脑数据的原始合成数据集,我们还表明,这种偏见与任务的难度有关。对这种偏见的认识对于开发可靠的,可推广的模型至关重要,这对于现实世界数据分布中的临床应用至关重要。

As interest in applying machine learning techniques for medical images continues to grow at a rapid pace, models are starting to be developed and deployed for clinical applications. In the clinical AI model development lifecycle (described by Lu et al. [1]), a crucial phase for machine learning scientists and clinicians is the proper design and collection of the data cohort. The ability to recognize various forms of biases and distribution shifts in the dataset is critical at this step. While it remains difficult to account for all potential sources of bias, techniques can be developed to identify specific types of bias in order to mitigate their impact. In this work we analyze how the distribution of scanner manufacturers in a dataset can contribute to the overall bias of deep learning models. We evaluate convolutional neural networks (CNN) for both classification and segmentation tasks, specifically two state-of-the-art models: ResNet [2] for classification and U-Net [3] for segmentation. We demonstrate that CNNs can learn to distinguish the imaging scanner manufacturer and that this bias can substantially impact model performance for both classification and segmentation tasks. By creating an original synthesis dataset of brain data mimicking the presence of more or less subtle lesions we also show that this bias is related to the difficulty of the task. Recognition of such bias is critical to develop robust, generalizable models that will be crucial for clinical applications in real-world data distributions.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源