论文标题
通过培训,卷积神经网络如何丢失空间信息
How deep convolutional neural networks lose spatial information with training
论文作者
论文摘要
机器学习的一个核心问题是,深网如何在高维度中学习任务。一个有吸引力的假设是,他们通过构建与任务无关的数据的代表来实现这一壮举。对于图像数据集,这种观察得到了以下观察,即在训练之后(而不是之前)训练后,随着信号通过网络传播,神经表示对在图像上作用的差异性的敏感性越来越不敏感。这种敏感性的丧失与性能相关,并且与训练过程中获得的白噪声的敏感性相关。这些事实无法解释,正如我们所证明的那样,当将白噪声添加到训练集的图像中时仍然存在。在这里,我们(i)在经验上显示了各种体系结构,即通过空间和频道汇总来实现图像差异性的稳定性,(ii)引入了模型量表检测任务,该任务重现了我们对空间汇总的经验观察,并且(iii)(iii)对与差异的敏感性相吻合,以及(iii)对差异和噪声的敏感性以及噪声量表的敏感性如何。发现量表取决于网络体系结构中的前进。我们发现,对噪声的敏感性提高是由于在由Relu单位纠正后,在合并过程中堆积的噪声堆积。
A central question of machine learning is how deep nets manage to learn tasks in high dimensions. An appealing hypothesis is that they achieve this feat by building a representation of the data where information irrelevant to the task is lost. For image datasets, this view is supported by the observation that after (and not before) training, the neural representation becomes less and less sensitive to diffeomorphisms acting on images as the signal propagates through the net. This loss of sensitivity correlates with performance, and surprisingly correlates with a gain of sensitivity to white noise acquired during training. These facts are unexplained, and as we demonstrate still hold when white noise is added to the images of the training set. Here, we (i) show empirically for various architectures that stability to image diffeomorphisms is achieved by both spatial and channel pooling, (ii) introduce a model scale-detection task which reproduces our empirical observations on spatial pooling and (iii) compute analitically how the sensitivity to diffeomorphisms and noise scales with depth due to spatial pooling. The scalings are found to depend on the presence of strides in the net architecture. We find that the increased sensitivity to noise is due to the perturbing noise piling up during pooling, after being rectified by ReLU units.