论文标题

有限的会员推理

Bounding Membership Inference

论文作者

Thudi, Anvith, Shumailov, Ilia, Boenisch, Franziska, Papernot, Nicolas

论文摘要

差异隐私(DP)是关于培训算法的隐私保证的事实上的标准。尽管经验观察DP降低了模型对现有成员推理(MI)攻击的脆弱性,但理论上的基础是文献中很大程度上缺少这种情况。在实践中,这意味着需要对模型进行DP培训,可以大大降低其准确性。 在本文中,当训练算法提供$(\ varepsilon,δ)$ -DP时,我们对任何MI对手的积极准确性(即攻击精度)提供了更严格的限制。我们的界限为新型隐私放大方案的设计提供了信息:在培训开始之前,有效的训练集是从较大的集合中进行的。我们发现这大大降低了MI阳性准确性。结果,我们的计划允许使用松散的DP保证来限制任何MI对手的成功;这样可以确保模型的准确性受到隐私保证的影响。尽管这显然有益于与他们需要训练更多的数据的实体,但它还可以改善学术文献中研究的基准测试的准确性权威权衡。因此,我们还发现,比在MNIST和CIFAR10上保证的DP保证更有效地降低了最先进的MI攻击(LIRA)的有效性。我们通过讨论MI绑定在机器学习领域的含义来结束。

Differential Privacy (DP) is the de facto standard for reasoning about the privacy guarantees of a training algorithm. Despite the empirical observation that DP reduces the vulnerability of models to existing membership inference (MI) attacks, a theoretical underpinning as to why this is the case is largely missing in the literature. In practice, this means that models need to be trained with DP guarantees that greatly decrease their accuracy. In this paper, we provide a tighter bound on the positive accuracy (i.e., attack precision) of any MI adversary when a training algorithm provides $(\varepsilon, δ)$-DP. Our bound informs the design of a novel privacy amplification scheme: an effective training set is sub-sampled from a larger set prior to the beginning of training. We find this greatly reduces the bound on MI positive accuracy. As a result, our scheme allows the use of looser DP guarantees to limit the success of any MI adversary; this ensures that the model's accuracy is less impacted by the privacy guarantee. While this clearly benefits entities working with far more data than they need to train on, it can also improve the accuracy-privacy trade-off on benchmarks studied in the academic literature. Consequently, we also find that subsampling decreases the effectiveness of a state-of-the-art MI attack (LiRA) much more effectively than training with stronger DP guarantees on MNIST and CIFAR10. We conclude by discussing implications of our MI bound on the field of machine unlearning.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源