论文标题
通过程序归一化的稳健而准确的作者归因
Robust and Accurate Authorship Attribution via Program Normalization
论文作者
论文摘要
由于深度学习的迅速发展,源代码归因方法已取得了显着的准确性。但是,最近的研究阐明了它们对对抗攻击的脆弱性。特别是,他们很容易被试图创造另一个作者伪造或掩盖原始作者的对手所欺骗。为了解决这些新出现的问题,我们将此安全挑战提出为一般威胁模型,即$ \ textit {关系对手} $,该模型允许在任何问题领域中应用任意数量的语义传播转换。我们的理论研究表明了鲁棒性和鲁棒性和准确性之间的权衡条件。在这些见解的激励下,我们提出了一个新颖的学习框架,$ \ textit {正常化和预测} $($ \ textit {n&p} $),从理论上讲,它保证了任何作者资格 - 归类方法的稳健性。我们对$ \ textit {n&p} $进行了广泛的评估,以捍卫针对最新的攻击方法的两种最新作者资格 - 属性方法。我们的评估表明,$ \ textit {n&p} $提高了对对抗性输入的准确性,比香草型号提高了多达70%。更重要的是,$ \ textit {n&p} $在运行速度超过40倍的同时,比对抗性训练提高了强大的精度至45%。
Source code attribution approaches have achieved remarkable accuracy thanks to the rapid advances in deep learning. However, recent studies shed light on their vulnerability to adversarial attacks. In particular, they can be easily deceived by adversaries who attempt to either create a forgery of another author or to mask the original author. To address these emerging issues, we formulate this security challenge into a general threat model, the $\textit{relational adversary}$, that allows an arbitrary number of the semantics-preserving transformations to be applied to an input in any problem space. Our theoretical investigation shows the conditions for robustness and the trade-off between robustness and accuracy in depth. Motivated by these insights, we present a novel learning framework, $\textit{normalize-and-predict}$ ($\textit{N&P}$), that in theory guarantees the robustness of any authorship-attribution approach. We conduct an extensive evaluation of $\textit{N&P}$ in defending two of the latest authorship-attribution approaches against state-of-the-art attack methods. Our evaluation demonstrates that $\textit{N&P}$ improves the accuracy on adversarial inputs by as much as 70% over the vanilla models. More importantly, $\textit{N&P}$ also increases robust accuracy to 45% higher than adversarial training while running over 40 times faster.