论文标题
通过全球标签推断和分类,强大的元代表性学习
Robust Meta-Representation Learning via Global Label Inference and Classification
论文作者
论文摘要
在元学习中,很少有射击学习(FSL)是一个核心问题,学习者必须有效地从几乎没有标记的示例中学习。在FSL中,特征预训练最近已成为一种越来越流行的策略,可以显着提高概括性能。但是,预训练的贡献经常被忽视和研究,理论上对其对元学习绩效的影响有限。此外,预训练需要一组一致的跨培训任务共享的全球标签,这在实践中可能不可用。在这项工作中,我们首先显示前培训和元学习之间的联系来解决上述问题。我们讨论为什么训练会产生更强大的元代表性,并将理论分析与现有作品和经验结果联系起来。其次,我们介绍了一种新颖的元学习算法Meta Label Learne(Mela),该算法通过跨任务推断全球标签来学习任务关系。这使我们能够为FSL利用预训练,即使全球标签不可用或定义不明。最后,我们介绍了一种增强的预训练程序,进一步改善了学到的元代表性。从经验上讲,梅拉(Mela)的表现优于各种基准的现有方法,尤其是在更具挑战性的环境下,培训任务的数量有限,并且标签是特定于任务的。我们还提供广泛的消融研究,以突出其关键特性。
Few-shot learning (FSL) is a central problem in meta-learning, where learners must efficiently learn from few labeled examples. Within FSL, feature pre-training has recently become an increasingly popular strategy to significantly improve generalization performance. However, the contribution of pre-training is often overlooked and understudied, with limited theoretical understanding of its impact on meta-learning performance. Further, pre-training requires a consistent set of global labels shared across training tasks, which may be unavailable in practice. In this work, we address the above issues by first showing the connection between pre-training and meta-learning. We discuss why pre-training yields more robust meta-representation and connect the theoretical analysis to existing works and empirical results. Secondly, we introduce Meta Label Learning (MeLa), a novel meta-learning algorithm that learns task relations by inferring global labels across tasks. This allows us to exploit pre-training for FSL even when global labels are unavailable or ill-defined. Lastly, we introduce an augmented pre-training procedure that further improves the learned meta-representation. Empirically, MeLa outperforms existing methods across a diverse range of benchmarks, in particular under a more challenging setting where the number of training tasks is limited and labels are task-specific. We also provide extensive ablation study to highlight its key properties.