Gradient-descent hardware-aware training and deployment for mixed-signal neuromorphic processors


Creative Commons License

Cakal U., Maryada M., Wu C., ULUSOY İ., Muir D. R.

Neuromorphic Computing and Engineering, cilt.4, sa.1, 2024 (ESCI) identifier identifier

  • Yayın Türü: Makale / Tam Makale
  • Cilt numarası: 4 Sayı: 1
  • Basım Tarihi: 2024
  • Doi Numarası: 10.1088/2634-4386/ad2ec3
  • Dergi Adı: Neuromorphic Computing and Engineering
  • Derginin Tarandığı İndeksler: Emerging Sources Citation Index (ESCI), Scopus
  • Anahtar Kelimeler: DYNAP-SE2, mixed-signal, neuromorphic, spiking neural networks
  • Orta Doğu Teknik Üniversitesi Adresli: Evet

Özet

Mixed-signal neuromorphic processors provide extremely low-power operation for edge inference workloads, taking advantage of sparse asynchronous computation within spiking neural networks (SNNs). However, deploying robust applications to these devices is complicated by limited controllability over analog hardware parameters, as well as unintended parameter and dynamical variations of analog circuits due to fabrication non-idealities. Here we demonstrate a novel methodology for offline training and deployment of SNNs to the mixed-signal neuromorphic processor DYNAP-SE2. Our methodology applies gradient-based training to a differentiable simulation of the mixed-signal device, coupled with an unsupervised weight quantization method to optimize the network’s parameters. Parameter noise injection during training provides robustness to the effects of quantization and device mismatch, making the method a promising candidate for real-world applications under hardware constraints and non-idealities. This work extends Rockpool, an open-source deep-learning library for SNNs, with support for accurate simulation of mixed-signal SNN dynamics. Our approach simplifies the development and deployment process for the neuromorphic community, making mixed-signal neuromorphic processors more accessible to researchers and developers.