Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci.
Sec. Neuromorphic Engineering
Volume 18 - 2024 | doi: 10.3389/fnins.2024.1372257
This article is part of the Research Topic Deep Spiking Neural Networks: Models, Algorithms and Applications View all 5 articles

Auto-Spikformer: Spikformer Architecture Search

Provisionally accepted
  • 1 Shenzhen Graduate School, Peking University, Shenzhen, Guangdong, China
  • 2 Peng Cheng Laboratory, Shenzhen, Guangdong Province, China
  • 3 Peking University, Beijing, Beijing Municipality, China
  • 4 Southern University of Science and Technology, Shenzhen, Guangdong Province, China

The final, formatted version of the article will be published soon.

    The integration of self-attention mechanisms into Spiking Neural Networks (SNNs) has garnered considerable interest in the realm of advanced deep learning, primarily due to their biological properties. Recent advancements in SNN architecture, such as Spikformer, have demonstrated promising outcomes. However, we observe that Spikformer may exhibit excessive energy consumption, potentially attributable to redundant channels and blocks. To mitigate this issue, we propose a one-shot Spiking Transformer Architecture Search method, namely Auto-Spikformer. \textcolor{blue}{Auto-Spikformer optimizes the transformer architecture using a weight entanglement method and the SNN inner parameters using the proposed Discrete Spiking Parameters Search (DSPS) methods.} Benefiting from these methods, the performance of subnets with weights inherited from the supernet without even retraining is comparable to the original Spikformer. Moreover, we propose a new fitness function aiming to find a Pareto optimal combination balancing energy consumption and accuracy. Our experimental results demonstrate the effectiveness of Auto-Spikformer, which outperforms the original Spikformer and most CNN or ViT models with even fewer parameters and lower energy consumption.

    Keywords: Spiking neural network (SNN), transformer, transformer architecture search, Network architecture search (NAS), Evolutionary Algorithm (EA)

    Received: 17 Jan 2024; Accepted: 05 Jul 2024.

    Copyright: © 2024 Che, Zhou, Niu, Ma, Fang, Chen, Shen, Yuan and Tian. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

    * Correspondence:
    Li Yuan, Shenzhen Graduate School, Peking University, Shenzhen, 518055, Guangdong, China
    Yonghong Tian, Shenzhen Graduate School, Peking University, Shenzhen, 518055, Guangdong, China

    Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.