论文标题
分配变化中对抗性鲁棒性的概括性
Generalizability of Adversarial Robustness Under Distribution Shifts
论文作者
论文摘要
经验和认证鲁棒性的最新进展有望提供可靠且可部署的深度神经网络(DNNS)。尽管取得了成功,但大多数现有的DNN鲁棒性评估都是对从训练模型的相同分布中采样的图像进行的。但是,在现实世界中,DNN可以部署在表现出重大分布变化的动态环境中。在这项工作中,我们迈出了第一步,朝着一方面彻底研究经验和认证的对抗性鲁棒性与另一方面的域概括之间的相互作用。为此,我们在多个领域上训练健壮的模型,并评估其在看不见的域上的准确性和鲁棒性。我们观察到:(1)经验和认证的鲁棒性都可以推广到看不见的域,并且(2)可推广性水平与输入视觉相似性不太相关,这是由源和目标域之间的FID衡量的。我们还扩展了研究以涵盖现实世界中的医学应用,在该应用中,对抗性增强显着提高了鲁棒性的概括,对清洁数据准确性的影响最小。
Recent progress in empirical and certified robustness promises to deliver reliable and deployable Deep Neural Networks (DNNs). Despite that success, most existing evaluations of DNN robustness have been done on images sampled from the same distribution on which the model was trained. However, in the real world, DNNs may be deployed in dynamic environments that exhibit significant distribution shifts. In this work, we take a first step towards thoroughly investigating the interplay between empirical and certified adversarial robustness on one hand and domain generalization on another. To do so, we train robust models on multiple domains and evaluate their accuracy and robustness on an unseen domain. We observe that: (1) both empirical and certified robustness generalize to unseen domains, and (2) the level of generalizability does not correlate well with input visual similarity, measured by the FID between source and target domains. We also extend our study to cover a real-world medical application, in which adversarial augmentation significantly boosts the generalization of robustness with minimal effect on clean data accuracy.