论文标题

围栏:用数据增强技术击败对抗性示例的平台

FenceBox: A Platform for Defeating Adversarial Examples with Data Augmentation Techniques

论文作者

Qiu, Han, Zeng, Yi, Zhang, Tianwei, Jiang, Yong, Qiu, Meikang

论文摘要

经过广泛研究,深度神经网络(DNN)容易受到对抗示例(AES)的影响。随着越来越高级的对抗攻击方法,已经设计了大量相应的防御解决方案,以增强DNN模型的鲁棒性。在推断之前,利用数据增强技术来进行预处理输入样本以消除对抗性扰动已成为一种普及。通过混淆DNN模型的梯度,这些方法可以击败相当多的常规攻击。不幸的是,引入了基于先进的基于梯度的攻击技术(例如BPDA和EOT),以使这些预处理效应无效。 在本文中,我们提出了栅栏,这是一个综合框架,旨在击败各种对抗性攻击。栅栏配备了来自三种不同类别的15种数据增强方法。我们全面评估这些方法可以有效地减轻各种对抗性攻击。栅栏还为用户提供了API,以便以不同的模式轻松地通过其模型部署防御:他们可以选择一种任意的预处理方法,或者即使在高级对抗攻击下,也可以选择功能的组合,以获得更好的鲁棒性保证。我们开源围栏,并希望它可以用作标准工具包,以促进对抗攻击和防御的研究。

It is extensively studied that Deep Neural Networks (DNNs) are vulnerable to Adversarial Examples (AEs). With more and more advanced adversarial attack methods have been developed, a quantity of corresponding defense solutions were designed to enhance the robustness of DNN models. It has become a popularity to leverage data augmentation techniques to preprocess input samples before inference to remove adversarial perturbations. By obfuscating the gradients of DNN models, these approaches can defeat a considerable number of conventional attacks. Unfortunately, advanced gradient-based attack techniques (e.g., BPDA and EOT) were introduced to invalidate these preprocessing effects. In this paper, we present FenceBox, a comprehensive framework to defeat various kinds of adversarial attacks. FenceBox is equipped with 15 data augmentation methods from three different categories. We comprehensively evaluated that these methods can effectively mitigate various adversarial attacks. FenceBox also provides APIs for users to easily deploy the defense over their models in different modes: they can either select an arbitrary preprocessing method, or a combination of functions for a better robustness guarantee, even under advanced adversarial attacks. We open-source FenceBox, and expect it can be used as a standard toolkit to facilitate the research of adversarial attacks and defenses.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源