论文标题
会员推断针对语义细分模型的攻击
Membership Inference Attacks Against Semantic Segmentation Models
论文作者
论文摘要
会员推理攻击旨在通过观察其预测来推断数据记录是否已使用数据记录来训练目标模型。在医疗保健等敏感领域中,这可能构成严重的侵犯隐私。在这项工作中,我们试图通过对语义图像分割领域中的成员推理攻击和防御进行详尽的研究来解决现有的知识差距。我们的发现表明,对于某些威胁模型,这些学习设置可能比以前考虑的分类设置更容易受到伤害。我们还研究了一个威胁模型,不诚实的对手可以执行模型中毒以帮助其推断并评估这些适应对成员推理攻击成功的影响。我们定量评估对各种语义分割任务中许多流行模型架构的攻击,这表明该领域中的会员推理攻击可以实现很高的成功率,并且对它们的辩护可能会导致不利的隐私性实用性权衡或增加计算成本。
Membership inference attacks aim to infer whether a data record has been used to train a target model by observing its predictions. In sensitive domains such as healthcare, this can constitute a severe privacy violation. In this work we attempt to address the existing knowledge gap by conducting an exhaustive study of membership inference attacks and defences in the domain of semantic image segmentation. Our findings indicate that for certain threat models, these learning settings can be considerably more vulnerable than the previously considered classification settings. We additionally investigate a threat model where a dishonest adversary can perform model poisoning to aid their inference and evaluate the effects that these adaptations have on the success of membership inference attacks. We quantitatively evaluate the attacks on a number of popular model architectures across a variety of semantic segmentation tasks, demonstrating that membership inference attacks in this domain can achieve a high success rate and defending against them may result in unfavourable privacy-utility trade-offs or increased computational costs.