Spike2Former: Efficient Spiking Transformer for High-performance Image Segmentation

Authors: Zhenxin Lei, Man Yao, Jiakui Hu, Xinhao Luo, Yanye Lu, Bo Xu, Guoqi Li

AAAI 2025 | Venue PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We set a new state-of-the-art for SNNs in various semantic segmentation datasets, with a significant improvement of +12.7% m Io U and 5.0 efficiency on ADE20K, +14.3% m Io U and 5.2 efficiency on VOC2012, and +9.1% m Io U and 6.6 efficiency on City Scapes.
Researcher Affiliation Academia 1University of Chinese Academy of Sciences 2Institute of Automation, Key Laboratory of Brain Cognition and Brain-inspired Intelligence Technology, Chinese Academy of Sciences 3Institute of Medical Technology, Peking University Health Science Center, Peking University 4National Biomedical Imaging Center, Peking University
Pseudocode No The paper describes methods through mathematical formulations and architectural diagrams (e.g., Figure 1, Equations 1-21) but does not include a dedicated pseudocode or algorithm block.
Open Source Code No The paper does not contain any explicit statement about releasing code, nor does it provide a link to a code repository.
Open Datasets Yes We conduct semantic segmentation on ADE20k (Zhou et al. 2017), City Scapes (Cordts et al. 2016), and Pascal VOC2012 (Everingham et al. 2010) datasets.
Dataset Splits No The paper mentions input sizes for datasets (e.g., 'ADE20k Input Size 512 512') and discusses training settings (e.g., 'Learning Rate', 'Optimizer', 'Training Steps'), but does not explicitly provide details about training, validation, or test dataset splits in the main text. It states 'More training details can be found in the Appendix.', but the appendix is not provided.
Hardware Specification No The paper does not provide specific hardware details (e.g., GPU models, CPU types, or memory specifications) used for running its experiments.
Software Dependencies No The paper mentions using 'Adam W' as an optimizer and 'Meta Spikeformer' as a backbone, but does not provide specific version numbers for any programming languages, libraries, or software frameworks used in the implementation or experimentation.
Experiment Setup Yes Table 2: Hyper-parameters setting in Spike2Former. Input Size 512 512 (ADE20k), 512 1024 (City Scapes), 512 512 (Pascal VOC2012). Learning Rate 2e-4 (ADE20k), 2e-3 (City Scapes), 2e-3 (Pascal VOC2012). Optimizer Adam W (all). Training Steps 160k (ADE20k), 90k (City Scapes), 80k (Pascal VOC2012).