论文标题

纠正决策树:探索可解释有效的机器学习的景观

Rectified Decision Trees: Exploring the Landscape of Interpretable and Effective Machine Learning

论文作者

Li, Yiming, Bai, Jiawang, Li, Jiawei, Yang, Xue, Jiang, Yong, Xia, Shu-Tao

论文摘要

解释性和有效性是现实中采用机器学习方法的两个必不可少的必不可少的要求。在本文中,我们提出了一个基于知识蒸馏的决策树扩展,称为纠正的决策树(REDT),以探索同时满足这些要求的可能性。具体而言,我们扩展了标准决策树的拆分标准和结束条件,该标准决策树的结束条件允许在保留确定性拆分路径的同时使用软标签进行训练。然后,我们根据从训练有素的教师模型中通过一种新颖的基于折刀的方法提取的软标签来训练REDT。因此,Redt保留了决策树的出色解释性,同时表现相对较好。在经验和理论上,采用软标签而不是硬标签的有效性也得到了分析。令人惊讶的是,实验表明,与标准决策树相比,软标签的引入也从总节点和规则的方面降低了模型的大小,这是从教师模型中提取的“黑暗知识”中意外的礼物。

Interpretability and effectiveness are two essential and indispensable requirements for adopting machine learning methods in reality. In this paper, we propose a knowledge distillation based decision trees extension, dubbed rectified decision trees (ReDT), to explore the possibility of fulfilling those requirements simultaneously. Specifically, we extend the splitting criteria and the ending condition of the standard decision trees, which allows training with soft labels while preserving the deterministic splitting paths. We then train the ReDT based on the soft label distilled from a well-trained teacher model through a novel jackknife-based method. Accordingly, ReDT preserves the excellent interpretable nature of the decision trees while having a relatively good performance. The effectiveness of adopting soft labels instead of hard ones is also analyzed empirically and theoretically. Surprisingly, experiments indicate that the introduction of soft labels also reduces the model size compared with the standard decision trees from the aspect of the total nodes and rules, which is an unexpected gift from the `dark knowledge' distilled from the teacher model.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源