Exploring the Sparsity-Quantization Interplay on a Novel Hybrid SNN Event-Driven Architecture
Abstract
Spiking Neural Networks (SNNs) offer potential advantages in energy efficiency but currently trail Artificial Neural Networks (ANNs) in versatility, largely due to challenges in efficient input encoding. Recent work shows that direct coding achieves superior accuracy with fewer timesteps than traditional rate coding. However, there is a lack of specialized hardware to fully exploit the potential of direct-coded SNNs, especially their mix of dense and sparse layers. This work proposes the first hybrid inference architecture for direct-coded SNNs. The proposed hardware architecture comprises a dense core to efficiently process the input layer and sparse cores optimized for event-driven spiking convolutions. Furthermore, for the first time, we investigate and quantify the quantization effect on sparsity. Our experiments on two variations of the VGG9 network and implemented on a Xilinx Virtex UltraScale+ FPGA (Field-Programmable Gate Array) reveal two novel findings. Firstly, quantization increases the network sparsity by up to with minimal loss of accuracy. Combined with the inherent low power benefits, this leads to a improvement in energy compared to the full-precision version. Secondly, direct coding outperforms rate coding, achieving a improvement in accuracy and consuming less energy per image. Overall, our accelerator achieves higher throughput and consumes half the power compared to previous work. Our accelerator code is available at: https://github.com/githubofaliyev/SNN-DSE/tree/DATE25.
Index Terms:
Spiking neural networks, sparsity-aware SNN, SNN accelerator, neuromorphic computing, quantization.I Introduction
Spiking Neural Networks (SNNs) offer a biologically inspired, energy-efficient alternative to traditional Artificial Neural Networks (ANNs) by utilizing sparse, event-driven computations. SNNs mimic the brain’s neurons by communicating through discrete pulses or ”spikes”. However, achieving accuracy comparable to ANNs, especially in deep SNNs, is challenging due to the complexities of encoding input data into spikes [1, 2]. Direct coding [3] is a promising approach that addresses this challenge by repeatedly presenting input samples over multiple timesteps, improving SNN accuracy with fewer timesteps compared to the popular rate coding approach [4]. In direct coding, the raw, floating-point input data is directly fed into the first convolution layer, which generates floating-point outputs. These floating-point outputs are processed by a spiking neuron layer that implements a threshold-based spike generation mechanism to produce the binary spikes that drive the rest of the SNN.
While prior work suggests that rate-coded networks may be more energy efficient due to the non-binary activations in direct-coded SNN input layers [4], we argue that the true energy-saving potential of direct coding remains largely untapped. Existing evaluations [4] are constrained by fixed timestep comparisons and platforms optimized for dense computations, overlooking the inherent benefits of variable sparse activity patterns characteristic of direct-coded SNNs. Moreover, traditional event-driven [5] or systolic array-based SNN implementations [6], while efficient for homogeneous computations, struggle to effectively handle the diverse layer characteristics present in direct-coded SNNs. This mismatch often leads to underutilization of processing elements in layers requiring fewer resources, further diminishing overall efficiency. To fully realize the potential of direct-coded SNNs, heterogeneous architectures with specialized hardware for both dense and sparse layers are essential.
In this paper, we propose the first-of-its-kind hybrid architecture explicitly designed for direct-coded SNNs. Our architecture leverages workload partitioning, assigning the input layer—characterized by the largest feature map dimensions, non-binary, and non-sparse activations—to a dedicated dense core. The remaining network layers, exhibiting varying degrees of sparsity, are processed by specialized sparse cores. Our hardware design is further distinguished by novel, parameterized, fully-pipelined, and high-throughput datapaths. We establish fine-grained, layer-wise workload models that encompass individual layer sizes, input feature map dimensions, input activation sparsity, and other key factors. These models serve as the foundation for design-time parameter selection, enabling the optimal partitioning of heterogeneous hardware resources to maximize performance. Furthermore, we conduct the first ablation study to investigate the impact of quantization on the intrinsic sparsity behavior of direct-coded SNN models, an important step toward realizing energy-efficient SNN implementations.
Our results on a Xilinx Virtex UltraScale+ FPGA across multiple datasets demonstrate the significant advantages of our hybrid architecture, quantization, and direct coding for SNNs. Quantization (with 4-bit integer weights and biases) substantially reduces spiking traffic, while direct coding achieves state-of-the-art accuracy with minimal timesteps, countering prior findings [4] and highlighting the potential for reduced hardware latency and improved energy efficiency. These results underscore the critical need for specialized architectures to fully unlock the efficiency potential of SNNs. Notably, compared to the most related work [7], our accelerator achieves lower power and higher throughput for a spiking VGG9 implementation on CIFAR100.

II Background
II-A Spiking Neuron Model
The spiking neuron model used in this work is a leaky integrate-and-fire (LIF) neuron [8], whose characteristics are shown in Equations 1 and 2.
(1) |
(2) |
Equation 1 describes how the neuron’s membrane potential () evolves over time. The decay factor (), ranging between and , controls how much the previous potential affects the current potential . A higher value implies less decay, enabling the neuron to retain more of its previous state, resulting in sparser behavior. Incoming weighted spikes from other neurons () increase the potential, while a threshold-based self-decay term () reduces it. Equation 2 determines when the neuron fires. If the membrane potential exceeds the threshold (), the neuron generates a spike (). As such a lower value reduces the potential required for firing, thereby increasing the neuron’s firing frequency.

II-B Quantization-Aware Training
We employ Quantization-Aware Training (QAT) [9] to quantize model weights and biases into integers, with the quantization error incorporated into the loss function during training. This enables the model to adapt to the quantization constraints, enhancing its performance when operating with quantized data. During evaluation, weights and biases are fully quantized, while neuronal parameters remain in floating-point due to the current lack of support for LIF neuron quantization. The accumulated membrane data undergoes dequantization back to floating-point for accurate spiking operations.
III Impacts of Quantization on Sparsity
The potential efficiency gains of low-bit quantization raise an intriguing question: can quantization influence the sparsity of an SNN model? To investigate this, we empirically compared a 32-bit floating-point network (called fp32) with its 4-bit integer (int4) counterpart, both trained using QAT. Section V details our experimental setup. Figure 1 illustrates the variability in the VGG9 model’s sparsity for SVHN, CIFAR10, and CIFAR100. Our experiments revealed two key insights:
-
1.
Comparable accuracy with both precisions: The int4 and fp32 networks demonstrate similar accuracy varying by only and for SVHN, CIFAR10, and CIFAR100, respectively.
-
2.
Quantization increases sparsity: Compared to fp32, int4 yields , , and fewer spikes for SVHN, CIFAR10, and CIFAR100, respectively. We hypothesize that introducing quantization noise during training implicitly encourages sparsity by selectively deactivating neurons, leading to a sparser and more efficient representation without sacrificing accuracy.
These findings suggest that quantization, beyond its traditional role in reducing model size and power consumption, may also induce sparsity. This opens exciting avenues for further exploring the synergy between quantization and sparsity-aware techniques to further enhance SNN efficiency.
IV Hybrid Hardware Architecture
Architecture Overview: Our hybrid architecture integrates specialized dense and sparse cores connected by on-chip FPGA memories. Core allocation follows a layer-wise strategy, with each core’s size tailored to the corresponding layer’s sparsity-driven workload needs. This layer-wise resource specialization is essential due to the potential for high sparsity variability across network layers. While our evaluation focuses on specific model architectures, the principles underlying this hybrid design—specialized processing cores and layer-wise resource allocation—can be readily adapted to support various SNN models and computational requirements.
IV-A Dense Core
Figure 2 depicts the block diagram of the dense core (dc), comprising three key components: a PE array, control units, and activation units. The PE array features a fixed column of 27 processing elements (PEs) and employs a weight stationary (WS) dataflow. We chose 27 PEs because of the WS requirements, specifically to handle input channels and filters. This choice minimizes memory footprint requirements, as the input layer’s three channels lead to fewer weights compared to input or output activations. Since we employ tiling in the output channel, we parameterize the number of rows, meaning that each row handles one output feature map. Each row then sequentially moves onto the next feature map until all output channels are processed. As a result, partial sums flow horizontally (left to right) while input image pixels flow vertically (top to down) in a systolic manner.
Each PE contains a Multiply and Accumulate (MAC) unit, along with a register for weight storage. All 27 PEs work in parallel to process input channels and generate a single output membrane potential per cycle. To maintain the systolic pipeline flow, we utilize independent registers/FFs to act as “shift registers” to provide a fixed delay for each of the 27 inputs in the array. These registers are controlled by the Staggering Routine within the Control unit. The right-most PE’s shift register depth dictates the array’s pipeline depth.
The Control unit is implemented as a state machine and orchestrates three main tasks:
-
•
Data management: It feeds image data into the PE array. The Address Generation Routine calculates read indices for the 27 top PEs, accessing image buffers (implemented with on-chip flip-flops for their small storage footprint). This enables row-major image storage while facilitating parallel weight access by all 27 PEs. The Staggering Routine manages ”shift registers” that control the data flow from the image buffers.
-
•
Activation synchronization: It starts/stops the Activ unit using the EN signal. Once the PE pipeline is filled (i.e., the first membrane potential accumulation is produced), the Control unit triggers the activation phase (responsible for spike train generation) with the signal EN.
-
•
Output channel tiling: The Control unit coordinates the tiling of output channels, meaning that each row in the systolic array strides through the output channels. When the rows of the systolic array transition to the next output feature map computation, it resets the partial sums and membrane potentials within the PE Array and Activ units using the rst signal. Finally, the unit sets the layer avail signal high to indicate the completion of layer processing.
The accumulated potential from each row flows into the Activ unit, where it performs two core functions. First, it modifies the incoming potential by adding the filter bias value, applying the leakage factor () to mimic the leaky behavior, and then performs thresholding. If the membrane potential exceeds the threshold (), the unit subtracts the threshold value from the membrane potential and sets the associated spike to ; otherwise, the spike is . Secondly, after processing an entire output feature map, the Activ unit writes the generated spike train to the Block RAM (BRAM), which serves as the intermediate storage between layers. Spike trains in BRAM follow a timestep-major order (see Figure 2) with the spike trains for consecutive timesteps stored in contiguous addresses. For example, if the layer consists of output channels and timesteps, then locations in total are required for the layer’s spike train storage. This layout is consistent for both dense and sparse cores.
IV-B Sparse Core
We adapt conventional convolution (CONV) and fully connected (FC) layers for event-driven processing by dividing their operation into spike train compression and accumulation phases. The sparse core design, illustrated in Figure 3, implements these phases and consists of two primary components: an Event Control Unit (ECU) and Neural Cores (NC). Within the ECU, the control routine is implemented as a state machine and has several key functions.
Compr. Routine: First, the ECU fetches a spike train from the input Spike RAM, and then tiles the whole spike train into -bit chunks, which are processed sequentially by the Compression (Compr.) routine. The core function of this routine is to eliminate the non-spiking (i.e., ’0’) bits from the bit array and generate a compact register array called (see Figure 3). This is achieved by processing bits per cycle and using a priority encoder to identify and send the address of the first set bit (’1’) within each chunk to the array, a common application of priority encoders, as shown in Figure 3 (right). Finally, the ECU’s bit reset component sets the identified ’1’ bit back to ’0’ in the previous cycle’s spike train, allowing the Compr. routine to locate the next set bit in subsequent cycles.

Address Generation: After compression, the accumulation phase begins, utilizing the register array. For each spike, there are 9 associated neurons in the feature map whose membrane potentials must be updated with corresponding filter coefficients. The Address Generation routine iterates through the filter coefficients and calculates the addresses of neurons associated with that spike event and filter size (e.g., a spike at affects the 9 neurons from to ). These signals are then directed to the neural cores (NCs). Importantly, the Compr. and accumulation phases execute in parallel, enabling the compression of new spike trains from subsequent input feature maps while simultaneously processing earlier feature map spike events.
With the pairs provided, the Accum routine within each NC performs the accumulation phase of the LIF neuron. It reads the membrane potential value from the BRAM, updates them with corresponding coefficient weights, and writes back the result to the BRAM. Similar to the dense core, the sparse cores also support quantization. Also, note that both the Address Generation and Accum routines are fully pipelined and can update one neuron per cycle. Filter weights for CONV layers are stored in BRAM and LUTRAM, while larger fully-connected (FC) weights use Ultra RAMs (URAMs) for their higher density and energy efficiency. After processing all input feature maps, the control unit enables the NC’s Activ. routine to initiate the LIF neuron’s spiking phase.
The architecture unrolls the output channels by a factor of , defined as a top-level parameter, to determine the number of NC instances. Each NC instance strides through the output channels by . For instance, an NC instance with index and will process OFMs with indices 1, 9, 17, etc. The architecture also features max-pooling on spikes, which aligns better with SNN temporal dynamics and minimizes information loss during downsampling, compared to other pooling methods [10]. Implementation on binary feature maps only requires sliding an OR gate over an input area, where is the downsampling ratio.
IV-C Memory Optimization
For efficient resource usage, we only use the FPGA’s on-chip memory (BRAM, URAM, LUTRAM) for storing model parameters and spike trains, avoiding the external DDR memory. Early layers with smaller convolutions use LUTRAM, which is more efficient than flip-flops for small RAMs due to its flexibility, despite being less abundant than flip-flops in Xilinx UltraScale+ devices.
Layer | LUT & FF | BRAM & URAM | Power* |
[W] | |||
int4 hardware | |||
CONV_1_1 | 1.9 & 1.9K | 0 & 0 | 0.048 |
CONV_1_2 | 11.7K & 14.6K | 32 & 0 | 0.205 |
CONV_2_1 | 1.7K & 2.1K | 44 & 0 | 0.054 |
CONV_2_2 | 5.1K & 5.1K | 164 & 0 | 0.17 |
CONV_3_1 | 1.6K & 1.3K | 144 & 0 | 0.1 |
CONV_3_2 | 5.7K & 5.2K | 216 & 0 | 0.293 |
CONV_3_3 | 5.8K & 5.1K | 211 & 0 | 0.284 |
FC | 6K & 2.1K | 168 & 0 | 0.125 |
Total | 109.7K & 37.6K | 979 & 0 | 1.231 |
Utilization | 6.43% & 1.11% | 30.23% & 0 | – |
fp32 hardware | |||
CONV_1_1 | 11.6K & 1.9K | 0 & 0 | 0.051 |
CONV_1_2 | 670.3K & 15.2K | 32 & 0 | 0.251 |
CONV_2_1 | 11.4K & 5.3K | 212 & 0 | 0.152 |
CONV_2_2 | 34.4K & 10.1K | 272 & 54 | 0.561 |
CONV_3_1 | 11.6K & 2.9K | 464 & 129 | 0.405 |
CONV_3_2 | 45.6K & 12.5K | 648 & 145 | 0.96 |
CONV_3_3 | 39.2K & 8.4K | 631 & 140 | 0.634 |
FC | 7.6K & 2.8K | 607 & 368 | 0.508 |
Total | 821.6K & 58.7K | 2466 & 836 | 3.471 |
Utilization | 47.73% & 1.73% | 91.2% & 65.69% | – |
*Instance-level dynamic power. Static power is and for int4 and fp32 respectively.
Our strategy for aggressive on-chip storage, while beneficial, comes with a trade-off in power consumption. Since a large portion of this storage remains inactive during layer execution (until NC reaches the next output channel), the associated weight data is not used. To address this, we implement clock gating for our memory units (right side of Figure 3). The memory is partitioned into two regions, with the most significant bit (MSB) of the address line controlling the active region for both reads and writes. An AND gate controls the clock signal to the memory unit, ensuring that only the active region receives clock cycles. While this on-chip storage strategy works well for quantized models like int4, it poses limitations for fp32 models. Specifically, the high LUT usage for weight storage prevents us from scaling beyond the custom VGG9 architecture or deploying larger models.



IV-D Quantization Support
Due to the lack of native quantization support, the design incorporates floating-point computation capabilities in both dense and sparse cores to de-quantize weights and biases retrieved from memory. We utilize a resource-efficient shift-and-add approach instead of DSP blocks to perform multiplications by constants [11], improving overall efficiency.
V Experimental Results
In this section, we first detail our experimental setup and leverage our novel hybrid architecture to analyze the benefits of quantization (resource utilization, energy, power, and latency). We then compare direct and rate coding, demonstrating the superiority of direct coding. Finally, we compare our hardware against two recent and relevant works.
V-A Experimental Setup
We implemented our proposed hardware using SystemVerilog and synthesized on a Xilinx Virtex® UltraScale+™ XCVU13P FPGA. For evaluation, we use the CIFAR-10, CIFAR-100, and Street View House Numbers (SVHN) datasets with a modified VGG9 network trained using snnTorch [12]. The network, trained using surrogate gradients [13], has a structure: 64C3-112C3-MP2-192C3-216C3-MP2-480C3-504C3-560C3-MP2-1064-. C denotes filters of size and MP denotes maxpooling. represents the number of neurons (i.e., population) in the network’s output layer. Prior work [14] has shown that using a population of neurons in the output layer can lead to better accuracy, even with fewer time steps. That is, we can make the output layer larger instead of needing more time steps to achieve good results. Through experimentation, we found that the following values provided the best accuracy with the minimum possible spike train length: for CIFAR10 and SVHN datasets, and for CIFAR100. We selected the network architecture to ensure compatibility with the Virtex UltraScale+ platform, following prior work [6]. We tuned the leaky neuron hyperparameters ( and ) and used layer-wise batch normalization to prevent overfitting.
Hardware configurations: For each dataset, we have one lightweight (LW) and two performance-optimized configurations (perf2 and perf4). The LW baseline prioritizes minimal resource usage while ensuring balanced layer-wise execution latency of the network. The two performance-optimized versions scale up resources by factors of and , respectively. To determine the optimal lightweight configuration, we modeled the trained network’s layer-wise workload distribution using Equation 3:
(3) |
where F is the number of filter coefficients (e.g., for ), C_out is the number of output channels, and is the number of spikes for input feature map . The spike information was acquired empirically by running the network once on the hardware. Our goal was to partition resources to minimize the execution latency difference between the most and least workload-intensive layers. All three configurations were synthesized to operate at a 100 MHz clock frequency.
V-B Resource Utilization
Table I summarizes the implementation results of our proposed hardware for CIFAR100 (in perf2), which are representative of the trends observedfor CIFAR10 and SVHN. We compare the results of full-precision (fp32) with its 4-bit integer counterpart (int4). We empirically determined that a configuration (neural cores allocated per layer) yields the most balanced execution profile (layer overheads: 0.9%, 13.4%, 13.6%, 13.8%, 12.8%, 12.3%, 12.9%, 15.6%, 4.8%). The int4 implementation demonstrates significant resource savings compared to fp32, using fewer LUTs, primarily due to efficient LUTRAM usage for the CONV_1_2 weight data. Additionally, int4 uses fewer BRAMs/URAMs due to the minimum 8-bit width configuration of BRAM primitives for CONV layers. Overall, the fp32 and int4 designs occupy and of the FPGA’s LUT resources, respectively. Importantly, the fp32 hardware consumes more power than the int4 configuration, highlighting the power efficiency benefits of quantization.
Coding | Time | Total | Acc. | Latency | Energy | Energy |
Steps | Spikes | [%] | [ms] | [mJ] | Imprv. | |
Rate | 25 | 107K | 77.37 | 340 | 201 | — |
Direct | 2 | 41K | 87.01 | 11.7 | 7.6 |
Dataset | Study | Network | Weight | Acc | Platform | Power | Latency1 | Energy1 | Throughput | |
Precision | [%] | [MHz] | [W] | [ms] | [mJ] | [FPS] | ||||
SVHN | [15] | VGG11 | 4-bit | 89 | ZCU102 | 200 | 0.4 | — | — | 65 |
perf4 | VGG9 | 93.9 | XCVU13P | 100 | 0.89 | 61 | 6.4 | 110 | ||
CIFAR10 | [15] | VGG11 | 4-bit | 78 | ZCU102 | 200 | 0.4 | — | — | 62 |
perf2 | VGG9 | 86.6 | XCVU13P | 100 | 0.73 | 59 | 4.9 | 120 | ||
CIFAR100 | [7] | VGG11 | 32-bit | 60.1 | XCVU13P | 115 | 4.9 | 210 | — | 4.7 |
perf4 | VGG9 | 56.9 | XCVU13P | 100 | 2.35 | 37 | 16.1 | 218 |
1’—’ indicates no reported results in prior work.
V-C Energy Analysis
Our energy analysis comparing int4 and fp32 (Figure 4) reveals significant energy savings benefits of quantization across CIFAR10, CIFAR100, and SVHN datasets. We calculate the energy expenditure per image by summing the energy per layer. For CIFAR10 and CIFAR100, int4 reduces the average energy by and , respectively, across all configurations. The majority of these savings (roughly for CIFAR10 and for CIFAR100) stem directly from int4’s inherent power advantage. Additionally, the fp32 designs exhibit a higher spike count ( and for CIFAR10 and CIFAR100, respectively), contributing a further to the energy difference. For SVHN, the scaling trend differs slightly due to the resource allocation dominance of the CONV_1_2 layer (with more allocated neural cores), leading to a smaller energy gap between fp32 and int4. The perf4 configuration of the quantized hardware achieves a energy reduction compared to its LW counterpart, contrasting with the reduction in the full-precision design. Since the full-precision hardware utilizes LUTRAM for this layer, increasing the NC count has a less pronounced impact on power compared to the quantized hardware. Overall, quantization yields a significant reduction in latency and notable energy savings across all datasets.
V-D Direct vs Rate coding comparison
Table II depicts our comparison of direct coding and rate coding on CIFAR10 using the quantized LW configuration. Because the rate-coded network receives spikes as inputs, it only needs sparse cores, while the direct-coded network needs our hybrid architecture. As such, although both networks run on our hardware architecture, for a fair comparison, we turned off the dense core for the rate-coded network to prevent unnecessary power consumption. The Total Spikes represents the spikes across all timesteps. With only 2 timesteps, direct coding yielded much higher sparsity (2.6 fewer spikes) and achieved higher accuracy than rate coding at 25 timesteps. Further increasing the timesteps plateaued the accuracy for both schemes111To enable reproducibility of our accuracy and implementation, our training and hardware code are available at https://github.com/githubofaliyev/SNN-DSE/tree/DATE25. Importantly, the direct-coded network consumed less energy than its rate-coded counterpart, contrary to prior work [3].
V-E Comparison to previous work
Table III compares our proposed hybrid architecture’s dynamic power and throughput to two recent related works: an event-driven design with quantization support [15] and a resource-efficient approach [7]. We focus on dynamic power and throughput as these metrics were reported in prior works.
Compared to [15], we achieve more than 2 the throughput for SVHN and CIFAR10, with a power increase of and , respectively (Table III). In addition to using a lower power board, [15] obtained lower power measurements by subtracting the board’s baseline power consumption from the accelerator’s total runtime power. Additionally, we outperform their work in inference accuracy ( and higher for SVHN and CIFAR10, respectively). Compared to [7], our perf4 configuration on CIFAR100 achieves throughput improvement and power savings, with only a accuracy decrease. This work is the most relevant comparison to ours due to the use of similar implementation platforms.
VI Conclusion
This paper presented a novel hybrid architecture that strategically combines dense and sparse cores, optimized for the unique workload characteristics of direct-coded SNNs. Our results demonstrate the significant advantages of direct coding, achieving both superior accuracy and hardware performance compared to traditional rate coding. We also highlight the benefits of quantization, showcasing its impact on sparsity and efficiency. The insights from this work lay the foundation for co-design approaches where SNN algorithms and hardware architectures are co-developed to maximize performance and enable more efficient SNN accelerators.
Future research will further explore direct coding through expanded evaluations with diverse models and datasets, including larger-scale networks that may require external memory access. While our current implementation focuses on models that fit within on-chip BRAM, additional studies are needed to analyze performance impacts when incorporating off-chip memory access for broader model support. This includes investigating techniques to minimize external memory bandwidth requirements and optimize the memory hierarchy for larger networks.
Acknowledgment
This work was partially supported by National Science Foundation Grant 1844952. Any opinions, findings, conclusions or recommendations expressed are those of the authors and do not necessarily reflect the views of the NSF.
References
- [1] W. Guo, M. E. Fouda, A. M. Eltawil, and K. N. Salama, “Neural coding in spiking neural networks: A comparative study for robust neuromorphic systems,” Frontiers in Neuroscience, vol. 15, p. 638474, 2021.
- [2] B. Rueckauer and S.-C. Liu, “Conversion of analog to spiking neural networks using sparse temporal coding,” in 2018 IEEE international symposium on circuits and systems (ISCAS), pp. 1–5, IEEE, 2018.
- [3] Y. Wu, L. Deng, G. Li, J. Zhu, Y. Xie, and L. Shi, “Direct training for spiking neural networks: Faster, larger, better,” in Proceedings of the AAAI conference on artificial intelligence, vol. 33, pp. 1311–1318, 2019.
- [4] Y. Kim, H. Park, A. Moitra, A. Bhattacharjee, Y. Venkatesha, and P. Panda, “Rate coding or direct coding: Which one is better for accurate, robust, and energy-efficient spiking neural networks?,” in ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 71–75, IEEE, 2022.
- [5] J. Sommer, M. A. Özkan, O. Keszocze, and J. Teich, “Efficient hardware acceleration of sparsely active convolutional spiking neural networks,” IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, vol. 41, no. 11, pp. 3767–3778, 2022.
- [6] M. T. L. Aung, D. Gerlinghoff, C. Qu, L. Yang, T. Huang, R. S. M. Goh, T. Luo, and W.-F. Wong, “Deepfire2: A convolutional spiking neural network accelerator on fpgas,” IEEE Transactions on Computers, vol. 72, no. 10, pp. 2847–2857, 2023.
- [7] D. Gerlinghoff, Z. Wang, X. Gu, R. S. M. Goh, and T. Luo, “A resource-efficient spiking neural network accelerator supporting emerging neural encoding,” in 2022 Design, Automation & Test in Europe Conference & Exhibition (DATE), pp. 92–95, IEEE, 2022.
- [8] P. Lansky, P. Sanda, and J. He, “The parameters of the stochastic leaky integrate-and-fire neuronal model,” Journal of Computational Neuroscience, vol. 21, pp. 211–223, 2006.
- [9] B. Jacob, S. Kligys, B. Chen, M. Zhu, M. Tang, A. Howard, H. Adam, and D. Kalenichenko, “Quantization and training of neural networks for efficient integer-arithmetic-only inference,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2704–2713, 2018.
- [10] W. Fang, Z. Yu, Y. Chen, T. Masquelier, T. Huang, and Y. Tian, “Incorporating learnable membrane time constant to enhance learning of spiking neural networks,” in Proceedings of the IEEE/CVF international conference on computer vision, pp. 2661–2671, 2021.
- [11] F. Dinechin, S. Filip, L. Forget, and M. Kumm, “Table-based versus shift-and-add constant multipliers for fpgas. 26th ieee symp,” in Computer Arithmetic, pp. 1–8, 2019.
- [12] J. K. Eshraghian, M. Ward, E. O. Neftci, X. Wang, G. Lenz, G. Dwivedi, M. Bennamoun, D. S. Jeong, and W. D. Lu, “Training spiking neural networks using lessons from deep learning,” Proceedings of the IEEE, 2023.
- [13] E. O. Neftci, H. Mostafa, and F. Zenke, “Surrogate gradient learning in spiking neural networks: Bringing the power of gradient-based optimization to spiking neural networks,” IEEE Signal Processing Magazine, vol. 36, no. 6, pp. 51–63, 2019.
- [14] I. Aliyev, K. Svoboda, and T. Adegbija, “Design space exploration of sparsity-aware application-specific spiking neural network accelerators,” IEEE Journal on Emerging and Selected Topics in Circuits and Systems, 2023.
- [15] S. Panchapakesan, Z. Fang, and J. Li, “Syncnn: Evaluating and accelerating spiking neural networks on fpgas,” ACM Transactions on Reconfigurable Technology and Systems, vol. 15, no. 4, pp. 1–27, 2022.