Complex Dynamic Neurons Improved Spiking Transformer Network for Efficient Automatic Speech Recognition
DOI:
https://doi.org/10.1609/aaai.v37i1.25081Keywords:
CMS: Brain Modeling, CMS: Agent & Cognitive Architectures, CMS: Simulating Humans, ML: Bio-Inspired LearningAbstract
The spiking neural network (SNN) using leaky-integrated-and-fire (LIF) neurons has been commonly used in automatic speech recognition (ASR) tasks. However, the LIF neuron is still relatively simple compared to that in the biological brain. Further research on more types of neurons with different scales of neuronal dynamics is necessary. Here we introduce four types of neuronal dynamics to post-process the sequential patterns generated from the spiking transformer to get the complex dynamic neuron improved spiking transformer neural network (DyTr-SNN). We found that the DyTr-SNN could handle the non-toy automatic speech recognition task well, representing a lower phoneme error rate, lower computational cost, and higher robustness. These results indicate that the further cooperation of SNNs and neural dynamics at the neuron and network scales might have much in store for the future, especially on the ASR tasks.Downloads
Published
2023-06-26
How to Cite
Wang, Q., Zhang, T., Han, M., Wang, Y., Zhang, D., & Xu, B. (2023). Complex Dynamic Neurons Improved Spiking Transformer Network for Efficient Automatic Speech Recognition. Proceedings of the AAAI Conference on Artificial Intelligence, 37(1), 102-109. https://doi.org/10.1609/aaai.v37i1.25081
Issue
Section
AAAI Technical Track on Cognitive Modeling & Cognitive Systems