论文标题

深神经网络不变,我们应该如何衡量它?

In What Ways Are Deep Neural Networks Invariant and How Should We Measure This?

论文作者

Kvinge, Henry, Emerson, Tegan H., Jorgenson, Grayson, Vasquez, Scott, Doster, Timothy, Lew, Jesse D.

论文摘要

通常说,深度学习模型对某些特定类型的转型是“不变的”。但是,该陈述的含义很大取决于制作的上下文。在本文中,我们探讨了深度学习模型的不变性和均衡性的性质,目的是更好地理解他们实际上在正式层面上捕获这些概念的方式。我们介绍了一个不变性和均衡度指标的家族,使我们能够以将它们与其他指标(例如损失或准确性)相关的方式量化这些属性。我们使用指标更好地理解用于建立网络不变性的两种最受欢迎​​的方法:数据增强和模棱两可的层。我们在深度学习模型中得出了一系列关于不变性和均衡性的结论,范围从预审预告量化的模型是否会影响训练有素的模型的不变性,以至于通过训练学习的不变性可以推广到分布数据。

It is often said that a deep learning model is "invariant" to some specific type of transformation. However, what is meant by this statement strongly depends on the context in which it is made. In this paper we explore the nature of invariance and equivariance of deep learning models with the goal of better understanding the ways in which they actually capture these concepts on a formal level. We introduce a family of invariance and equivariance metrics that allows us to quantify these properties in a way that disentangles them from other metrics such as loss or accuracy. We use our metrics to better understand the two most popular methods used to build invariance into networks: data augmentation and equivariant layers. We draw a range of conclusions about invariance and equivariance in deep learning models, ranging from whether initializing a model with pretrained weights has an effect on a trained model's invariance, to the extent to which invariance learned via training can generalize to out-of-distribution data.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源