TY - JOUR
T1 - Improving SpikeProp’s Training Efficiency in Spiking Neural Networks for Large Language Models Through Innovative Weight Initialization
AU - Ahmed, Falah. Y. H.
AU - Zakarya, Muhammad
AU - Khan, Naveed
AU - Zebari, Dilovan Asaad
AU - Al-Bahri, Mahmood
AU - Joseph, Bwalya Kelvin
AU - Abdullah, Abdullah
N1 - © The Author(s) 2025.
PY - 2025/11/6
Y1 - 2025/11/6
N2 - Spiking neural networks (SNNs) use individual temporal spikes for computation and communication, simulating the actions of biological neurons. SNN had long been disregarded since it was thought to be intricate and difficult to analyze. We investigate the improvement of SpikeProp, a supervised learning model tailored for SNNs, in this work. Three distinct models are being proposed and investigated, including the proposed model 1, the proposed model 2, and the proposed model 3, each providing unique improvements to the SpikeProp algorithm. To accelerate convergence and adaptive learning rates, particle swarm optimization (PSO) and momentum factors are integrated into the proposed model 1. In proposed model 2, a rate dependency is introduced based on angle-driven learning. By incorporating PSO and learning rates, model 3 combines the strengths of both models 1 and 2. We believe, SNNs can be trained and classified more efficiently and accurately using these models. Furthermore, we examine how large language models (LLMs) might inform the design and interpretability of neural architectures and learning methodologies while also enhancing SNN training. Through the use of LLMs, we seek to enhance model transparency and encourage more Responsible AI (RAI) principles. A thorough evaluation and comparison of proposed models with traditional methods confirms that these models consistently outperform traditional methods for various real datasets. Consequently, they have a high potential for practical applications in neural network training in real-world settings and LLM-informed development, contributing to the advancement of AI systems.
AB - Spiking neural networks (SNNs) use individual temporal spikes for computation and communication, simulating the actions of biological neurons. SNN had long been disregarded since it was thought to be intricate and difficult to analyze. We investigate the improvement of SpikeProp, a supervised learning model tailored for SNNs, in this work. Three distinct models are being proposed and investigated, including the proposed model 1, the proposed model 2, and the proposed model 3, each providing unique improvements to the SpikeProp algorithm. To accelerate convergence and adaptive learning rates, particle swarm optimization (PSO) and momentum factors are integrated into the proposed model 1. In proposed model 2, a rate dependency is introduced based on angle-driven learning. By incorporating PSO and learning rates, model 3 combines the strengths of both models 1 and 2. We believe, SNNs can be trained and classified more efficiently and accurately using these models. Furthermore, we examine how large language models (LLMs) might inform the design and interpretability of neural architectures and learning methodologies while also enhancing SNN training. Through the use of LLMs, we seek to enhance model transparency and encourage more Responsible AI (RAI) principles. A thorough evaluation and comparison of proposed models with traditional methods confirms that these models consistently outperform traditional methods for various real datasets. Consequently, they have a high potential for practical applications in neural network training in real-world settings and LLM-informed development, contributing to the advancement of AI systems.
KW - Spiking neural network
KW - Classification
KW - Angel-driven dependency
KW - Particle swarm optimization
UR - https://www.scopus.com/pages/publications/105021084735
U2 - 10.1007/s44196-025-00961-x
DO - 10.1007/s44196-025-00961-x
M3 - Article
SN - 1875-6891
VL - 18
SP - 1
EP - 26
JO - International Journal of Computational Intelligence Systems
JF - International Journal of Computational Intelligence Systems
IS - 1
M1 - 286
ER -