论文标题
通过混合转换和尖峰计时依赖反向传播,使深度尖峰神经网络能够
Enabling Deep Spiking Neural Networks with Hybrid Conversion and Spike Timing Dependent Backpropagation
论文作者
论文摘要
尖刺神经网络(SNN)使用异步离散事件(或尖峰)运行,这可能会导致神经形态硬件实现的能源效率更高。许多作品表明,可以通过复制训练有素的人工神经网络(ANN)并将每一层的点火阈值作为该层中接收到的最大输入来形成推理的SNN。这些类型的转换后的SNN需要大量的时间步骤才能达到竞争精度,从而减少了节能。通过从头开始训练基于尖峰的反向传播的SNN可以减少时间步骤的数量,但这在计算上昂贵且缓慢。为了应对这些挑战,我们为深SNN提供了一种计算效率的培训技术。我们提出了一种混合训练方法:1)进行转换后的SNN,并将其权重和阈值作为基于尖峰的反向流磁的初始化步骤,以及2)在此精心初始化的网络上执行相关的依赖性峰值依赖性依赖性反向流动(STDB),以在几个epochs中获得几个频率的SNN,并需要在几个epochs中进行汇总,并需要几乎几步。 STDB使用新型的替代梯度函数,该功能使用神经元的尖峰时间定义。所提出的培训方法在不到20个时期的基于尖峰的反向传播中收敛于大多数标准图像分类数据集,从而大大降低了训练的复杂性,而训练的复杂性与SCRATCH的训练SNN相比。我们在CIFAR-10,CIFAR-100和Imagenet数据集上对VGG和Resnet架构进行实验。我们在SNN上使用250个时间步骤的ImageNet数据集获得了65.19%的TOP-1精度,与以相似精度相似的转换SNN相比,它的速度快10倍。
Spiking Neural Networks (SNNs) operate with asynchronous discrete events (or spikes) which can potentially lead to higher energy-efficiency in neuromorphic hardware implementations. Many works have shown that an SNN for inference can be formed by copying the weights from a trained Artificial Neural Network (ANN) and setting the firing threshold for each layer as the maximum input received in that layer. These type of converted SNNs require a large number of time steps to achieve competitive accuracy which diminishes the energy savings. The number of time steps can be reduced by training SNNs with spike-based backpropagation from scratch, but that is computationally expensive and slow. To address these challenges, we present a computationally-efficient training technique for deep SNNs. We propose a hybrid training methodology: 1) take a converted SNN and use its weights and thresholds as an initialization step for spike-based backpropagation, and 2) perform incremental spike-timing dependent backpropagation (STDB) on this carefully initialized network to obtain an SNN that converges within few epochs and requires fewer time steps for input processing. STDB is performed with a novel surrogate gradient function defined using neuron's spike time. The proposed training methodology converges in less than 20 epochs of spike-based backpropagation for most standard image classification datasets, thereby greatly reducing the training complexity compared to training SNNs from scratch. We perform experiments on CIFAR-10, CIFAR-100, and ImageNet datasets for both VGG and ResNet architectures. We achieve top-1 accuracy of 65.19% for ImageNet dataset on SNN with 250 time steps, which is 10X faster compared to converted SNNs with similar accuracy.