论文标题
早期学习的正则化阻止了嘈杂标签的记忆
Early-Learning Regularization Prevents Memorization of Noisy Labels
论文作者
论文摘要
我们提出了一个新颖的框架,以在嘈杂注释的情况下通过深度学习进行分类。在嘈杂的标签上接受培训时,已经观察到深层神经网络在“早期学习”阶段首先使用干净的标签拟合训练数据,然后最终用错误的标签记住示例。我们证明,即使在简单的线性模型中,早期的学习和记忆也是高维分类任务中的基本现象,并且在这种情况下给出了理论上的解释。在这些发现的激励下,我们开发了一种针对嘈杂分类任务的新技术,该技术利用了早期学习阶段的进步。与现有方法相反,与现有的方法相比,在早期学习过程中使用模型输出来检测示例,并忽略或试图纠正错误标签,我们采取了不同的路线,而是通过正则化来利用早期学习。我们的方法有两个关键要素。首先,我们利用半监督的学习技术来基于模型输出产生目标概率。其次,我们设计一个正规化术语,将模型转向这些目标,隐含地阻止了错误标签的记忆。结果框架显示出对几个标准基准和现实世界数据集的嘈杂注释的鲁棒性,在该基准和现实世界数据集中取得了与艺术状态相当的结果。
We propose a novel framework to perform classification via deep learning in the presence of noisy annotations. When trained on noisy labels, deep neural networks have been observed to first fit the training data with clean labels during an "early learning" phase, before eventually memorizing the examples with false labels. We prove that early learning and memorization are fundamental phenomena in high-dimensional classification tasks, even in simple linear models, and give a theoretical explanation in this setting. Motivated by these findings, we develop a new technique for noisy classification tasks, which exploits the progress of the early learning phase. In contrast with existing approaches, which use the model output during early learning to detect the examples with clean labels, and either ignore or attempt to correct the false labels, we take a different route and instead capitalize on early learning via regularization. There are two key elements to our approach. First, we leverage semi-supervised learning techniques to produce target probabilities based on the model outputs. Second, we design a regularization term that steers the model towards these targets, implicitly preventing memorization of the false labels. The resulting framework is shown to provide robustness to noisy annotations on several standard benchmarks and real-world datasets, where it achieves results comparable to the state of the art.