论文标题

超出模型压缩中的分类精度指标

Going Beyond Classification Accuracy Metrics in Model Compression

论文作者

Joseph, Vinu, Siddiqui, Shoaib Ahmed, Bhaskara, Aditya, Gopalakrishnan, Ganesh, Muralidharan, Saurav, Garland, Michael, Ahmed, Sheraz, Dengel, Andreas

论文摘要

随着边缘计算设备的增加,部署能源和资源有效模型的需求越来越大。大量研究专门用于开发可以大大减少模型大小的方法,而不会影响标准指标,例如TOP-1准确性。但是,这些修剪方法往往会导致其他指标的不匹配,例如跨阶层的公平性和解释性。为了打击这种未对准,我们提出了一种受知识依据文献启发的新型多部分损失功能。通过广泛的实验,我们证明了方法在不同的压缩算法,体系结构,任务以及数据集中的有效性。特别是,我们在压缩模型和参考模型之间的预测不匹配数量中最多可降低$ 4.1 \ times $ $,如果参考模型可以正确预测,则最高可获得$ 5.7 \ times $。同时不对压缩算法进行任何更改,也没有对损耗函数进行少量修改。此外,我们证明了模型预测之间的简单对齐方式自然可以改善其他指标的一致性,包括公平和归因。因此,我们的框架可以用作将来的简单插件组件,用于压缩算法。

With the rise in edge-computing devices, there has been an increasing demand to deploy energy and resource-efficient models. A large body of research has been devoted to developing methods that can reduce the size of the model considerably without affecting the standard metrics such as top-1 accuracy. However, these pruning approaches tend to result in a significant mismatch in other metrics such as fairness across classes and explainability. To combat such misalignment, we propose a novel multi-part loss function inspired by the knowledge-distillation literature. Through extensive experiments, we demonstrate the effectiveness of our approach across different compression algorithms, architectures, tasks as well as datasets. In particular, we obtain up to $4.1\times$ reduction in the number of prediction mismatches between the compressed and reference models, and up to $5.7\times$ in cases where the reference model makes the correct prediction; all while making no changes to the compression algorithm, and minor modifications to the loss function. Furthermore, we demonstrate how inducing simple alignment between the predictions of the models naturally improves the alignment on other metrics including fairness and attributions. Our framework can thus serve as a simple plug-and-play component for compression algorithms in the future.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源