论文标题
对抗性鲁棒的成像网模型会更好地传递吗?
Do Adversarially Robust ImageNet Models Transfer Better?
论文作者
论文摘要
转移学习是深度学习中广泛使用的范式,在该范式中,可以有效地适应标准数据集预先训练的模型。通常,更好的预训练模型会产生更好的转移结果,这表明初始准确性是转移学习绩效的关键方面。在这项工作中,我们确定了另一个这样的方面:我们发现,对抗性的模型虽然不准确,但在用于转移学习时的表现往往比其标准训练的同行更好。具体而言,我们专注于对抗性鲁棒的成像网分类器,并表明它们在标准的下游分类任务上的准确性提高了。进一步的分析在转移学习的背景下发现了健壮模型和标准模型之间的更多差异。我们的结果与最近的假设相一致(实际上是),表明鲁棒性会导致提高特征表示。我们的代码和模型可在https://github.com/microsoft/robust-models-transfer上找到。
Transfer learning is a widely-used paradigm in deep learning, where models pre-trained on standard datasets can be efficiently adapted to downstream tasks. Typically, better pre-trained models yield better transfer results, suggesting that initial accuracy is a key aspect of transfer learning performance. In this work, we identify another such aspect: we find that adversarially robust models, while less accurate, often perform better than their standard-trained counterparts when used for transfer learning. Specifically, we focus on adversarially robust ImageNet classifiers, and show that they yield improved accuracy on a standard suite of downstream classification tasks. Further analysis uncovers more differences between robust and standard models in the context of transfer learning. Our results are consistent with (and in fact, add to) recent hypotheses stating that robustness leads to improved feature representations. Our code and models are available at https://github.com/Microsoft/robust-models-transfer .