论文标题
用于移动应用程序的描述符模型的压缩
Compression of descriptor models for mobile applications
论文作者
论文摘要
深层神经网络通过新的大型和多样化的数据集出现了基于功能的图像匹配的最新性能。但是,在评估这些模型的计算成本,模型大小和匹配的准确性权衡方面,几乎没有工作。本文通过考虑最先进的硬核模型来明确解决这些实用指标。我们观察到学到的权重的显着冗余,我们通过使用深度可分离层和有效的塔克分解来利用它们。我们证明,这些方法的组合非常有效,但仍然牺牲了高端准确性。为了解决这一问题,我们提出了深度深度(CDP)层的卷积,该层提供了一种在标准和深度可分离卷积之间插值的手段。借助该提出的层,我们可以减少硬核模型的参数数量的8倍,计算复杂性降低13倍,同时牺牲跨hpatchesbenchs的整体准确性少于1%。为了进一步证明这种方法的概括,我们将其应用于最先进的超级点模型,在那里我们可以显着减少参数和浮点操作的数量,并且在匹配的准确性中最小的降解。
Deep neural networks have demonstrated state-of-the-art performance for feature-based image matching through the advent of new large and diverse datasets. However, there has been little work on evaluating the computational cost, model size, and matching accuracy tradeoffs for these models. This paper explicitly addresses these practical metrics by considering the state-of-the-art HardNet model. We observe a significant redundancy in the learned weights, which we exploit through the use of depthwise separable layers and an efficient Tucker decomposition. We demonstrate that a combination of these methods is very effective, but still sacrifices the top-end accuracy. To resolve this, we propose the Convolution-Depthwise-Pointwise(CDP) layer, which provides a means of interpolating between the standard and depthwise separable convolutions. With this proposed layer, we can achieve an 8 times reduction in the number of parameters on the HardNet model, 13 times reduction in the computational complexity, while sacrificing less than 1% on the overall accuracy across theHPatchesbenchmarks. To further demonstrate the generalisation of this approach, we apply it to the state-of-the-art SuperPoint model, where we can significantly reduce the number of parameters and floating-point operations, with minimal degradation in the matching accuracy.