论文标题

关于深度学习不确定性工具包的调查

A Survey on Uncertainty Toolkits for Deep Learning

论文作者

Pintz, Maximilian, Sicking, Joachim, Poretschkin, Maximilian, Akila, Maram

论文摘要

深度学习(DL)的成功促进了创建统一框架(例如Tensorflow或Pytorch)的创建,就像由其创造所驱动的。具有共同的构件有助于促进例如模型或概念的交换,并使开发更容易复制。尽管如此,DL模型的强大和可靠的评估和评估通常已被证明具有挑战性。这与他们的安全性越来越多,最近在“值得信赖的ML”领域达到顶峰。我们认为,除其他外,还进一步统一了根据工具包(即小型且专业的框架衍生品)来维护方法论,可能会对可信度和可重复性的问题产生积极影响。为此,我们介绍了DL中关于不确定性估计(UE)工具包的首次调查,因为UE在评估模型可靠性方面构成了基石。我们研究了11个有关建模和评估功能的工具包,为三个最有前途的三个最有前途的比较提供了深入的比较,即Pyro,Tensorflow概率和不确定性量化360。尽管前两个提供了很大的灵活性和无缝集成到各自的框架中,但最后一个具有较大的方法论范围。

The success of deep learning (DL) fostered the creation of unifying frameworks such as tensorflow or pytorch as much as it was driven by their creation in return. Having common building blocks facilitates the exchange of, e.g., models or concepts and makes developments easier replicable. Nonetheless, robust and reliable evaluation and assessment of DL models has often proven challenging. This is at odds with their increasing safety relevance, which recently culminated in the field of "trustworthy ML". We believe that, among others, further unification of evaluation and safeguarding methodologies in terms of toolkits, i.e., small and specialized framework derivatives, might positively impact problems of trustworthiness as well as reproducibility. To this end, we present the first survey on toolkits for uncertainty estimation (UE) in DL, as UE forms a cornerstone in assessing model reliability. We investigate 11 toolkits with respect to modeling and evaluation capabilities, providing an in-depth comparison for the three most promising ones, namely Pyro, Tensorflow Probability, and Uncertainty Quantification 360. While the first two provide a large degree of flexibility and seamless integration into their respective framework, the last one has the larger methodological scope.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源