论文标题
关于卷积神经网络的鲁棒性和可传递性
On Robustness and Transferability of Convolutional Neural Networks
论文作者
论文摘要
现代深层卷积网络(CNN)通常因不在分配转变下概括而受到批评。但是,转移学习中最近的一些突破表明,这些网络可以应对严重的分配变化,并成功地适应了一些培训示例中的新任务。在这项工作中,我们首次研究了现代图像分类CNN的分布和转移性能之间的相互作用,并研究了培训前数据大小,模型量表和数据预处理管道的影响。我们发现,增加训练集和模型大小都显着改善了分布转移的鲁棒性。此外,我们表明,可能令人惊讶的是,在预处理中的简单更改(例如修改图像分辨率)在某些情况下可以大大减轻鲁棒性问题。最后,我们概述了现有鲁棒性评估数据集的缺点,并引入了一个合成数据集SI分数,我们用于跨视觉数据中常见的变异因素(例如对象大小和位置)进行系统分析。
Modern deep convolutional networks (CNNs) are often criticized for not generalizing under distributional shifts. However, several recent breakthroughs in transfer learning suggest that these networks can cope with severe distribution shifts and successfully adapt to new tasks from a few training examples. In this work we study the interplay between out-of-distribution and transfer performance of modern image classification CNNs for the first time and investigate the impact of the pre-training data size, the model scale, and the data preprocessing pipeline. We find that increasing both the training set and model sizes significantly improve the distributional shift robustness. Furthermore, we show that, perhaps surprisingly, simple changes in the preprocessing such as modifying the image resolution can significantly mitigate robustness issues in some cases. Finally, we outline the shortcomings of existing robustness evaluation datasets and introduce a synthetic dataset SI-Score we use for a systematic analysis across factors of variation common in visual data such as object size and position.