SOTAVerified

Benchmarking Deep Spiking Neural Networks on Neuromorphic Hardware

2020-04-03Code Available0· sign in to hype

Christoph Ostrau, Jonas Homburg, Christian Klarhorst, Michael Thies, Ulrich Rückert

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

With more and more event-based neuromorphic hardware systems being developed at universities and in industry, there is a growing need for assessing their performance with domain specific measures. In this work, we use the methodology of converting pre-trained non-spiking to spiking neural networks to evaluate the performance loss and measure the energy-per-inference for three neuromorphic hardware systems (BrainScaleS, Spikey, SpiNNaker) and common simulation frameworks for CPU (NEST) and CPU/GPU (GeNN). For analog hardware we further apply a re-training technique known as hardware-in-the-loop training to cope with device mismatch. This analysis is performed for five different networks, including three networks that have been found by an automated optimization with a neural architecture search framework. We demonstrate that the conversion loss is usually below one percent for digital implementations, and moderately higher for analog systems with the benefit of much lower energy-per-inference costs.

Tasks

Reproductions