论文标题
通过Langevin Dynamics学习和推断稀疏编码模型
Learning and Inference in Sparse Coding Models with Langevin Dynamics
论文作者
论文摘要
我们描述了一个随机,动态的系统,能够在概率潜在变量模型中进行推理和学习。在此类模型中,最具挑战性的问题 - 对潜在变量的后验分布进行采样 - 提议通过利用电子和神经系统固有的自然来源来解决。我们通过得出通过Langevin Dynamics推断其潜在变量的连续时间方程来证明稀疏编码模型的想法。模型参数是通过根据另一个连续时间方程式同时进化来学习的,从而绕开了对数字蓄能器或全局时钟的需求。此外,我们表明兰格文动力学导致了从“ L0稀疏”制度中后部分布进行采样的有效过程,在该方案中鼓励潜在变量设置为零,而不是具有小的L1标准。这使该模型可以正确地融合了稀疏性概念,而不必诉诸放松的稀疏性以进行优化。对合成和自然图像数据集的提议动力系统的仿真表明,该模型能够概率地正确推断,从而可以学习词典以及先前的参数。
We describe a stochastic, dynamical system capable of inference and learning in a probabilistic latent variable model. The most challenging problem in such models - sampling the posterior distribution over latent variables - is proposed to be solved by harnessing natural sources of stochasticity inherent in electronic and neural systems. We demonstrate this idea for a sparse coding model by deriving a continuous-time equation for inferring its latent variables via Langevin dynamics. The model parameters are learned by simultaneously evolving according to another continuous-time equation, thus bypassing the need for digital accumulators or a global clock. Moreover we show that Langevin dynamics lead to an efficient procedure for sampling from the posterior distribution in the 'L0 sparse' regime, where latent variables are encouraged to be set to zero as opposed to having a small L1 norm. This allows the model to properly incorporate the notion of sparsity rather than having to resort to a relaxed version of sparsity to make optimization tractable. Simulations of the proposed dynamical system on both synthetic and natural image datasets demonstrate that the model is capable of probabilistically correct inference, enabling learning of the dictionary as well as parameters of the prior.