论文标题
分布式学习的强大而有效的聚合
Robust and Efficient Aggregation for Distributed Learning
论文作者
论文摘要
分布式学习范式(例如联合和分散的学习)允许在代理集合中协调模型,而无需交换原始数据。取而代之的是,代理根据其可用数据计算本地更新,然后与参数服务器或其同行共享更新模型。接下来是一个聚合步骤,传统上采用(加权)平均值的形式。已知基于平均的分布式学习方案容易受到异常值的影响。单个恶意代理能够将基于平均的分布式学习算法推向任意差的模型。这促使了强大的聚合方案的发展,这些方案基于中位数和修剪平均值的变化。尽管此类程序确保了对异常值和恶意行为的稳健性,但它们的成本大大降低了样本效率。这意味着,当前强大的聚合方案需要明显更高的代理参与率,以达到给定的性能水平,而不是其在非污染设置中的平均基准。在这项工作中,我们通过为分布式学习开发统计高效且强大的聚合方案来纠正这一缺点。
Distributed learning paradigms, such as federated and decentralized learning, allow for the coordination of models across a collection of agents, and without the need to exchange raw data. Instead, agents compute model updates locally based on their available data, and subsequently share the update model with a parameter server or their peers. This is followed by an aggregation step, which traditionally takes the form of a (weighted) average. Distributed learning schemes based on averaging are known to be susceptible to outliers. A single malicious agent is able to drive an averaging-based distributed learning algorithm to an arbitrarily poor model. This has motivated the development of robust aggregation schemes, which are based on variations of the median and trimmed mean. While such procedures ensure robustness to outliers and malicious behavior, they come at the cost of significantly reduced sample efficiency. This means that current robust aggregation schemes require significantly higher agent participation rates to achieve a given level of performance than their mean-based counterparts in non-contaminated settings. In this work we remedy this drawback by developing statistically efficient and robust aggregation schemes for distributed learning.