论文标题
Hoyer正规器是超长延长峰值神经网络所需的全部
Hoyer regularizer is all you need for ultra low-latency spiking neural networks
论文作者
论文摘要
尖峰神经网络(SNN)已成为一种有吸引力的时空计算范式,用于广泛的低功率视觉任务。但是,最新的(SOTA)SNN模型要么产生多个时间步骤,从而阻碍了他们在实时用例中的部署,要么显着提高培训复杂性。为了减轻这种关注,我们为一次性阶段SNN提供了一个培训框架(从头开始),该框架使用了最近提出的Hoyer正常使用器的新型变体。我们估计每个SNN层的阈值是其激活图的剪裁版本的Hoyer极值,在该版本中,使用Hoyer正常器梯度下降训练了剪裁阈值。这种方法不仅降低了可训练的阈值的值,从而发出了大量的重量更新峰值,以有限的迭代次数(仅由于一个时间步长),还可以将膜电位值从阈值中移开,从而减轻噪声的效果,从而降低SNN准确度的噪声效果。我们的方法在复杂的图像识别任务方面的精确度权衡方面优于现有的尖峰,二进制和加法神经网络。对象检测的下游实验也证明了我们方法的功效。
Spiking Neural networks (SNN) have emerged as an attractive spatio-temporal computing paradigm for a wide range of low-power vision tasks. However, state-of-the-art (SOTA) SNN models either incur multiple time steps which hinder their deployment in real-time use cases or increase the training complexity significantly. To mitigate this concern, we present a training framework (from scratch) for one-time-step SNNs that uses a novel variant of the recently proposed Hoyer regularizer. We estimate the threshold of each SNN layer as the Hoyer extremum of a clipped version of its activation map, where the clipping threshold is trained using gradient descent with our Hoyer regularizer. This approach not only downscales the value of the trainable threshold, thereby emitting a large number of spikes for weight update with a limited number of iterations (due to only one time step) but also shifts the membrane potential values away from the threshold, thereby mitigating the effect of noise that can degrade the SNN accuracy. Our approach outperforms existing spiking, binary, and adder neural networks in terms of the accuracy-FLOPs trade-off for complex image recognition tasks. Downstream experiments on object detection also demonstrate the efficacy of our approach.