论文标题
STN:通过结构感知训练和自适应压缩的可扩展张量化网络
STN: Scalable Tensorizing Networks via Structure-Aware Training and Adaptive Compression
论文作者
论文摘要
深度神经网络(DNNS)在许多计算机视觉的任务中都表现出色。但是,流行体系结构的过度参数化表示会大大提高其计算复杂性和存储成本,并阻止其在边缘设备中的可用性,并受到约束资源。不管许多张量分解(TD)方法已通过压缩DNN来学习紧凑的表示,他们在实践中遭受了不可忽略的性能降低。在本文中,我们提出了可扩展的张力网络(STN),该网络会动态和自适应地调整模型大小和分解结构而无需重新培训。首先,我们通过添加低级别的正规器来确保网络所需的低级别特性以全张量格式来考虑训练期间的压缩。然后,考虑到网络层表现出各种低级别结构,STN是通过数据驱动的自适应TD方法获得的,从预训练的模型中学习了每个层分解的拓扑结构,并且在指定的存储约束下适当地选择了等级。结果,STN与任意网络体系结构兼容,并且在其他张力版本上实现了更高的压缩性能和灵活性。关于几种流行架构和基准测试的全面实验证实了我们模型对提高参数效率的优势。
Deep neural networks (DNNs) have delivered a remarkable performance in many tasks of computer vision. However, over-parameterized representations of popular architectures dramatically increase their computational complexity and storage costs, and hinder their availability in edge devices with constrained resources. Regardless of many tensor decomposition (TD) methods that have been well-studied for compressing DNNs to learn compact representations, they suffer from non-negligible performance degradation in practice. In this paper, we propose Scalable Tensorizing Networks (STN), which dynamically and adaptively adjust the model size and decomposition structure without retraining. First, we account for compression during training by adding a low-rank regularizer to guarantee networks' desired low-rank characteristics in full tensor format. Then, considering network layers exhibit various low-rank structures, STN is obtained by a data-driven adaptive TD approach, for which the topological structure of decomposition per layer is learned from the pre-trained model, and the ranks are selected appropriately under specified storage constraints. As a result, STN is compatible with arbitrary network architectures and achieves higher compression performance and flexibility over other tensorizing versions. Comprehensive experiments on several popular architectures and benchmarks substantiate the superiority of our model towards improving parameter efficiency.