论文标题
稀疏本地线性模型的神经发电机,用于实现准确性和可解释性
Neural Generators of Sparse Local Linear Models for Achieving both Accuracy and Interpretability
论文作者
论文摘要
对于可靠性,重要的是,机器学习方法的预测是人类可以解释的。通常,深层神经网络(DNN)可以提供准确的预测,尽管很难解释为什么DNN获得了这种预测。另一方面,线性模型的解释很容易,尽管它们的预测性能将很低,因为现实世界中的数据通常是本质上非线性的。为了将DNN高预测性能的好处和线性模型的高解释性结合到单个模型中,我们提出了稀疏局部线性模型(NGSLLS)的神经发生器。稀疏的局部线性模型具有很高的灵活性,因为它们可以近似非线性函数。 NGSLL使用采用每个样本的原始表示(例如单词序列)及其简化表示(例如,词袋)作为输入的DNN生成稀疏的线性权重。通过从原始表示形式中提取功能,权重可以包含丰富的信息以实现高预测性能。此外,该预测是可以解释的,因为它是通过简化表示和稀疏权重之间的内部产物获得的,其中只有我们的门模块在NGSLL中仅选择了少量权重。在使用现实世界数据集的实验中,我们通过评估预测性能和可视化图像和文本分类任务上生成的权重,来证明NGSLL的有效性。
For reliability, it is important that the predictions made by machine learning methods are interpretable by human. In general, deep neural networks (DNNs) can provide accurate predictions, although it is difficult to interpret why such predictions are obtained by DNNs. On the other hand, interpretation of linear models is easy, although their predictive performance would be low since real-world data is often intrinsically non-linear. To combine both the benefits of the high predictive performance of DNNs and high interpretability of linear models into a single model, we propose neural generators of sparse local linear models (NGSLLs). The sparse local linear models have high flexibility as they can approximate non-linear functions. The NGSLL generates sparse linear weights for each sample using DNNs that take original representations of each sample (e.g., word sequence) and their simplified representations (e.g., bag-of-words) as input. By extracting features from the original representations, the weights can contain rich information to achieve high predictive performance. Additionally, the prediction is interpretable because it is obtained by the inner product between the simplified representations and the sparse weights, where only a small number of weights are selected by our gate module in the NGSLL. In experiments with real-world datasets, we demonstrate the effectiveness of the NGSLL quantitatively and qualitatively by evaluating prediction performance and visualizing generated weights on image and text classification tasks.