论文标题
差异私有联合线性匪徒
Differentially-Private Federated Linear Bandits
论文作者
论文摘要
分散学习系统的快速扩散要求需要差异私人合作学习。在本文中,我们在上下文线性匪徒的上下文中研究了这一点:我们考虑合作解决共同上下文强盗的代理集合,同时确保他们的交流保持私密。对于这个问题,我们设计了一种用于集中式和分散的联合学习的多种私人算法。我们在遗憾中对其实用性进行了严格的技术分析,改善了合作匪徒学习的几个结果,并提供了严格的隐私保证。我们的算法在各种多代理设置中都在伪界和经验基准性能方面提供了竞争性能。
The rapid proliferation of decentralized learning systems mandates the need for differentially-private cooperative learning. In this paper, we study this in context of the contextual linear bandit: we consider a collection of agents cooperating to solve a common contextual bandit, while ensuring that their communication remains private. For this problem, we devise \textsc{FedUCB}, a multiagent private algorithm for both centralized and decentralized (peer-to-peer) federated learning. We provide a rigorous technical analysis of its utility in terms of regret, improving several results in cooperative bandit learning, and provide rigorous privacy guarantees as well. Our algorithms provide competitive performance both in terms of pseudoregret bounds and empirical benchmark performance in various multi-agent settings.