Abstract
Brain-inspired Spiking Neural Networks (SNNs) leverage sparse spikes to represent information and process them in an asynchronous event-driven manner, offering an energy-efficient paradigm for the next generation of machine intelligence. However, the current focus within the SNN community prioritizes accuracy optimization through the development of large-scale models, limiting their viability in resource-constrained and low-power edge devices. To address this challenge, we introduce a lightweight and hardware-friendly Quantized SNN (Q-SNN) that applies quantization to both synaptic weights and membrane potentials. By significantly compressing these two key elements, the proposed Q-SNNs substantially reduce both memory usage and computational complexity. Moreover, to prevent the performance degradation caused by this compression, we present a new Weight-Spike Dual Regulation (WS-DR) method inspired by information entropy theory. Experimental evaluations on various datasets, including static and neuromorphic, demonstrate that our Q-SNNs outperform existing methods in terms of both model size and accuracy. These state-of-the-art results in efficiency and efficacy suggest that the proposed method can significantly improve edge intelligent computing.
Original language | English |
---|---|
Title of host publication | 32nd ACM Multimedia Conference (MM 2024), 28 October - 1 November 2024, Melbourne, Australia |
Publisher | ACM |
Pages | 1-10 |
Number of pages | 10 |
Publication status | Accepted/In press - 20 Jul 2024 |
Event | 32nd ACM Multimedia Conference (MM 2024) - Melbourne, Australia Duration: 28 Oct 2024 → 1 Nov 2024 Conference number: 32 https://2024.acmmm.org/ |
Conference
Conference | 32nd ACM Multimedia Conference (MM 2024) |
---|---|
Country/Territory | Australia |
City | Melbourne |
Period | 28/10/24 → 1/11/24 |
Internet address |
Keywords
- Spiking Neural Networks,
- Neuromorphic Datasets
- Quantization