论文标题

通过减少培训开销的算法和MMV问题的算法正在展开

Algorithm Unfolding for Block-sparse and MMV Problems with Reduced Training Overhead

论文作者

Hauffen, Jan Christian, Jung, Peter, Mücke, Nicole

论文摘要

在本文中,我们认为在只有少量训练样本的情况下,多重测量向量(MMV)问题的算法正在展开。算法展开已显示出以数据驱动的方式凭经验加速的各种经典迭代算法的收敛性,但是对于监督学习,使用最小的培训数据实现这一目标很重要。为此,我们认为在不同的培训策略下,学到的障碍迭代收缩阈值算法(LBISTA)。要在最小训练开销上几乎无数据优化,必须大大减少算法展开的可训练参数的数量。因此,我们根据MMV观察模型施加的Kronecker结构明确提出了一个减少大小的网络体系结构,并在这种情况下介绍了相应的理论。为了确保适当的概括,我们将Lui等人的分析权重方法扩展到LBISTA和MMV设置。为这种情况指出了严格的理论保证和收敛结果。我们表明,可以通过在降低的MMV尺寸上求解明确方程来计算网络权重,该方程也可以接收封闭形式的解决方案。然后,我们考虑了卷积观察模型,并表明可以进一步简化提出的体系结构和分析权重计算,从而为卷积神经网络开辟了新的方向。最后,我们评估了数值实验中展开的算法,并讨论了与其他稀疏恢复算法的连接。

In this paper we consider algorithm unfolding for the Multiple Measurement Vector (MMV) problem in the case where only few training samples are available. Algorithm unfolding has been shown to empirically speed-up in a data-driven way the convergence of various classical iterative algorithms but for supervised learning it is important to achieve this with minimal training data. For this we consider learned block iterative shrinkage thresholding algorithm (LBISTA) under different training strategies. To approach almost data-free optimization at minimal training overhead the number of trainable parameters for algorithm unfolding has to be substantially reduced. We therefore explicitly propose a reduced-size network architecture based on the Kronecker structure imposed by the MMV observation model and present the corresponding theory in this context. To ensure proper generalization, we then extend the analytic weight approach by Lui et al to LBISTA and the MMV setting. Rigorous theoretical guarantees and convergence results are stated for this case. We show that the network weights can be computed by solving an explicit equation at the reduced MMV dimensions which also admits a closed-form solution. Towards more practical problems, we then consider convolutional observation models and show that the proposed architecture and the analytical weight computation can be further simplified and thus open new directions for convolutional neural networks. Finally, we evaluate the unfolded algorithms in numerical experiments and discuss connections to other sparse recovering algorithms.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源