论文标题
学习聚合功能
Learning Aggregation Functions
论文作者
论文摘要
由于其广泛的适用性,因此在机器学习社区中越来越多地在集合中学习。通常,通过使用固定的聚合函数(例如总和或最大值)来计算集合的表示形式。但是,最近的结果表明,通过总和(或最大)分解的通用函数表示需要高度不连续的(因此可学习的)映射,或者的潜在维度等于集合中最大元素数量。为了减轻此问题,我们引入了一个可学习的聚合功能(LAF),以进行任意基数。 LAF可以近似几个广泛使用的聚合器(例如平均,总和,最大)以及更复杂的功能(例如,方差和偏度)。我们报告了关于半合成和实际数据的实验,表明LAF胜过最先进的分解体系结构,例如深处集和基于图书馆的架构,例如主要邻域聚集,并且可以有效地与基于注意力的建筑结构相结合。
Learning on sets is increasingly gaining attention in the machine learning community, due to its widespread applicability. Typically, representations over sets are computed by using fixed aggregation functions such as sum or maximum. However, recent results showed that universal function representation by sum- (or max-) decomposition requires either highly discontinuous (and thus poorly learnable) mappings, or a latent dimension equal to the maximum number of elements in the set. To mitigate this problem, we introduce a learnable aggregation function (LAF) for sets of arbitrary cardinality. LAF can approximate several extensively used aggregators (such as average, sum, maximum) as well as more complex functions (e.g., variance and skewness). We report experiments on semi-synthetic and real data showing that LAF outperforms state-of-the-art sum- (max-) decomposition architectures such as DeepSets and library-based architectures like Principal Neighborhood Aggregation, and can be effectively combined with attention-based architectures.