论文标题
科学机器学习中推断外推的可解释模型
Interpretable models for extrapolation in scientific machine learning
论文作者
论文摘要
数据驱动的模型对于科学发现至关重要。为了实现最先进的模型准确性,研究人员正在采用越来越复杂的机器学习算法,这些算法通常超过了插值设置中的简单回归(例如随机K折线交叉验证),但外推性能,便携性和人类可解释性却遭受了较差的限制,这限制了它们促进新型科学科学洞察力的潜力。在这里,我们研究了广泛的科学和工程问题的模型性能与可解释性之间的权衡,并着重于材料科学数据集。我们将黑匣子随机森林和神经网络机器学习算法的性能与单功能线性回归的性能进行了比较,这些算法使用简单的随机搜索算法发现的可解释的输入功能拟合。对于插值问题,线性回归的平均预测误差的高度是黑匣子模型的平均预测误差。值得注意的是,当预测任务需要外推时,线性模型仅产生的平均误差仅比黑匣子模型高5%,并且在大约40%的经过测试的预测任务中超过了黑匣子模型,这表明它们在许多挤压问题中可能在复杂的算法上是可取的,因为它们具有出色的解释性,计算性的高架,并使用了。结果挑战了一个普遍的假设,即科学机器学习的外推模型受到性能和可解释性之间的固有权衡的约束。
Data-driven models are central to scientific discovery. In efforts to achieve state-of-the-art model accuracy, researchers are employing increasingly complex machine learning algorithms that often outperform simple regressions in interpolative settings (e.g. random k-fold cross-validation) but suffer from poor extrapolation performance, portability, and human interpretability, which limits their potential for facilitating novel scientific insight. Here we examine the trade-off between model performance and interpretability across a broad range of science and engineering problems with an emphasis on materials science datasets. We compare the performance of black box random forest and neural network machine learning algorithms to that of single-feature linear regressions which are fitted using interpretable input features discovered by a simple random search algorithm. For interpolation problems, the average prediction errors of linear regressions were twice as high as those of black box models. Remarkably, when prediction tasks required extrapolation, linear models yielded average error only 5% higher than that of black box models, and outperformed black box models in roughly 40% of the tested prediction tasks, which suggests that they may be desirable over complex algorithms in many extrapolation problems because of their superior interpretability, computational overhead, and ease of use. The results challenge the common assumption that extrapolative models for scientific machine learning are constrained by an inherent trade-off between performance and interpretability.