论文标题
Optilime:针对诊断计算机算法的优化石灰说明
OptiLIME: Optimized LIME Explanations for Diagnostic Computer Algorithms
论文作者
论文摘要
局部可解释的模型不足解释(LIME)是对任何类型的机器学习(ML)模型的解释性的流行方法。它一次通过学习围绕预测的简单线性模型来解释一个ML预测。该模型是在随机生成的数据点上训练的,从训练数据集分布中取样,并根据参考点的距离进行加权 - 石灰正在解释的。功能选择仅用于保留最重要的变量。石灰在不同领域的普遍存在,尽管其不稳定性(单个预测可能会得到不同的解释)是主要缺点之一。这是由于采样步骤中的随机性以及调整权重的灵活性并确定在检索到的解释中缺乏可靠性,从而使石灰采用有问题。尤其是在医学上,考虑到危及的决定的重要性和相关法律问题,临床专业人员的信任是必须确定可解释算法的接受。在本文中,我们重点介绍了解释的稳定性和依从性之间的权衡,即它与ML模型相似。利用我们的创新发现,我们提出了一个框架,以最大程度地提高稳定性,同时保留预定义的遵守水平。 Optilime提供了选择最佳依从性权衡水平的自由,更重要的是,它显然强调了检索到的解释的数学特性。结果,根据目前的问题,为从业者提供了工具来决定解释是否可靠。我们在玩具数据集上进行了广泛的测试 - 以视觉上的几何发现和医学数据集进行了观察。在后者中,我们从医学和数学的角度展示了该方法如何提出有意义的解释。
Local Interpretable Model-Agnostic Explanations (LIME) is a popular method to perform interpretability of any kind of Machine Learning (ML) model. It explains one ML prediction at a time, by learning a simple linear model around the prediction. The model is trained on randomly generated data points, sampled from the training dataset distribution and weighted according to the distance from the reference point - the one being explained by LIME. Feature selection is applied to keep only the most important variables. LIME is widespread across different domains, although its instability - a single prediction may obtain different explanations - is one of the major shortcomings. This is due to the randomness in the sampling step, as well as to the flexibility in tuning the weights and determines a lack of reliability in the retrieved explanations, making LIME adoption problematic. In Medicine especially, clinical professionals trust is mandatory to determine the acceptance of an explainable algorithm, considering the importance of the decisions at stake and the related legal issues. In this paper, we highlight a trade-off between explanation's stability and adherence, namely how much it resembles the ML model. Exploiting our innovative discovery, we propose a framework to maximise stability, while retaining a predefined level of adherence. OptiLIME provides freedom to choose the best adherence-stability trade-off level and more importantly, it clearly highlights the mathematical properties of the retrieved explanation. As a result, the practitioner is provided with tools to decide whether the explanation is reliable, according to the problem at hand. We extensively test OptiLIME on a toy dataset - to present visually the geometrical findings - and a medical dataset. In the latter, we show how the method comes up with meaningful explanations both from a medical and mathematical standpoint.