论文标题

对抗核心选择可进行有效的稳健训练

Adversarial Coreset Selection for Efficient Robust Training

论文作者

Dolatabadi, Hadi M., Erfani, Sarah, Leckie, Christopher

论文摘要

神经网络容易受到对抗性攻击的攻击:在其输入中添加精心设计的,不可察觉的扰动可以改变其输出。对抗性训练是针对此类攻击进行训练强大模型的最有效方法之一。不幸的是,这种方法比神经网络的香草培训要慢得多,因为它需要在每次迭代时为整个培训数据构建对抗性示例。通过利用核心选择理论,我们展示了选择一小部分训练数据的方式如何提供一种有原则的方法来降低稳健训练的时间复杂性。为此,我们首先为对抗核心选择提供融合保证。特别是,我们表明,收敛界限直接与我们的核心在整个训练数据上计算的梯度近似的效果直接相关。在我们的理论分析中,我们建议使用此梯度近似误差作为对抗核心选择目标,以有效地减少训练设置的大小。构建后,我们将在培训数据的这一子集上进行对抗训练。与现有方法不同,我们的方法可以适应各种培训目标,包括交易,$ \ ell_p $ -PGD和感知对抗性培训。我们进行了广泛的实验,以证明我们的进近可以使对抗性训练加快2-3倍,同时在清洁和稳健的精度中略有降解。

Neural networks are vulnerable to adversarial attacks: adding well-crafted, imperceptible perturbations to their input can modify their output. Adversarial training is one of the most effective approaches to training robust models against such attacks. Unfortunately, this method is much slower than vanilla training of neural networks since it needs to construct adversarial examples for the entire training data at every iteration. By leveraging the theory of coreset selection, we show how selecting a small subset of training data provides a principled approach to reducing the time complexity of robust training. To this end, we first provide convergence guarantees for adversarial coreset selection. In particular, we show that the convergence bound is directly related to how well our coresets can approximate the gradient computed over the entire training data. Motivated by our theoretical analysis, we propose using this gradient approximation error as our adversarial coreset selection objective to reduce the training set size effectively. Once built, we run adversarial training over this subset of the training data. Unlike existing methods, our approach can be adapted to a wide variety of training objectives, including TRADES, $\ell_p$-PGD, and Perceptual Adversarial Training. We conduct extensive experiments to demonstrate that our approach speeds up adversarial training by 2-3 times while experiencing a slight degradation in the clean and robust accuracy.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源