论文标题
使用熵损失进行健壮的深度学习
Towards Robust Deep Learning using Entropic Losses
论文作者
论文摘要
当前的深度学习解决方案是因为不通知他们是否可以在推断期间可靠地对示例进行分类而闻名。建立更可靠的深度学习解决方案的最有效方法之一是在所谓的分布式检测任务中提高其性能,该任务本质上包括“知道您不知道”或“知道未知”。换句话说,当不培训神经网络的课程实例时,分布外检测能力的系统可能会拒绝执行废话分类。本文通过提出新的损失功能和检测分数来解决挑衅的分布检测任务。不确定性估计也是建立更强大的深度学习系统的关键辅助任务。因此,我们还处理了与鲁棒性相关的任务,该任务评估了深神经网络所表现出的概率的现实性。为了证明我们的方法的有效性,除了包括最新结果的大量实验外,我们还基于最大熵原理来建立所提出方法的理论基础。与大多数当前方法不同,我们的损失和分数是无缝且原则性的解决方案,除了快速有效的推断外,还可以产生准确的预测。此外,我们的方法只需替换训练深神网络并计算快速检测的损失即可将我们的方法纳入当前和未来的项目中。
Current deep learning solutions are well known for not informing whether they can reliably classify an example during inference. One of the most effective ways to build more reliable deep learning solutions is to improve their performance in the so-called out-of-distribution detection task, which essentially consists of "know that you do not know" or "know the unknown". In other words, out-of-distribution detection capable systems may reject performing a nonsense classification when submitted to instances of classes on which the neural network was not trained. This thesis tackles the defiant out-of-distribution detection task by proposing novel loss functions and detection scores. Uncertainty estimation is also a crucial auxiliary task in building more robust deep learning systems. Therefore, we also deal with this robustness-related task, which evaluates how realistic the probabilities presented by the deep neural network are. To demonstrate the effectiveness of our approach, in addition to a substantial set of experiments, which includes state-of-the-art results, we use arguments based on the principle of maximum entropy to establish the theoretical foundation of the proposed approaches. Unlike most current methods, our losses and scores are seamless and principled solutions that produce accurate predictions in addition to fast and efficient inference. Moreover, our approaches can be incorporated into current and future projects simply by replacing the loss used to train the deep neural network and computing a rapid score for detection.