The paper explores the properties of spiking lottery tickets (SLTs) and compares them to standard lottery tickets (LTs) in both convolutional neural network (CNN) and transformer-based spiking neural network (SNN) structures.
For CNN-based models, the authors find that inner SLTs achieve higher sparsity and fewer performance losses compared to LTs (Reward 1). For transformer-based models, SLTs incur less accuracy loss compared to LTs counterparts at the same level of multi-level sparsity (Reward 2).
The authors propose a multi-level sparsity exploring algorithm for spiking transformers, which effectively achieves sparsity in the patch embedding projection (ConvPEP) module's weights, activations, and input patch numbers. Extensive experiments on RGB and event-based datasets demonstrate that the proposed SLT methods outperform standard LTs while achieving extreme energy savings (>80.0%).
The paper also analyzes the impact of spiking neural network parameters, such as time step and decay rate, on the performance of SLTs. The results show that increasing the time step can improve the performance of SLTs, while the optimal decay rate exhibits a non-monotonic relationship.
Іншою мовою
із вихідного контенту
arxiv.org
Ключові висновки, отримані з
by Hao Cheng,Ji... о arxiv.org 03-29-2024
https://arxiv.org/pdf/2309.13302.pdfГлибші Запити