Edge-AI & Neuromorphic – Efficient Processing of Time Series Data for Control & Prediction
Klaus Knobloch a
a Infineon Technologies Dresden, Koenigsbruecker Str. 180, Dresden, Germany
Proceedings of Materials, devices and systems for neuromorphic computing 2022 (MatNeC22)
Groningen, Netherlands, 2022 March 28th - 29th
Organizers: Jasper van der Velde, Elisabetta Chicca, Yoeri van de Burgt and Beatriz Noheda
Invited Speaker, Klaus Knobloch, presentation 003
DOI: https://doi.org/10.29363/nanoge.matnec.2022.003
Publication date: 23rd February 2022

Driven by mobility and internet-of-things, edge-AI has become increasingly important. Whether it is home applications, autonomous driving or fields like sensor networks and drones, all benefit from local AI data processing close to the device or sensor.

Thus, largely reducing energy consumption avoiding high-bandwidth data transport. Autonomous driving or drones doesn’t even allow communication with bigger processing infrastructure due to real-time latency requirements.

In areas where personal data like audio, visual or even medical data is processed, it is a requirement for privacy and security to handle the data as much as possible local on the device.

However, todays AI algorithms require intensive data processing limiting the scope of edge-AI applications. Although impressive progress has been reached by technology scaling to run GPU, NPU hardware accelerators also on mobile devices, still the power consumption remains a problem to be solved. For level 5 autonomous robot taxis the overall data processing would consume appr. 30% of the battery range of today’s electric cars [1].

And even such edge-AI is still focused on “static” classification of data, e.g. images or “signatures over time”. In contrast, AI processing of time-series sensor data provides further challenges. Prediction of movement of objects in radar requires history to identify only the relevant echoes (KI-ASIC) Systems depending on complex chemical and physical properties, like batteries [2], drones or car electric drives [3], need to include them in the control & prediction loop.

Recurrent neural networks (RNN) or long-short term memory (LSTM) neural networks are applied [4] but have strong disadvantages in multiply and accumulate acceleration due to complex neuron structure of input, forget and output gates as well as limited data re-use [5]. Neuromorphic, spiking neural networks (SNN) are much more hardware friendly working in the digital domain [6], but still lacking good supervised training algorithms. Here new concepts are available like surrogate gradients [7] to allow for backpropagation training.

For a radar gesture recognition, it is shown that such training can be done with backpropagation through time.  Efficiently small networks with only 10...30 hidden leaky integrate-and-fire neurons are on-par or even able to outperform LSTM networks. The network is implemented on a digital simulation SpiNNaker 2 FPGA prototype in a real time closed loop system. Impact of quantization and buffer memory usage is investigated to optimize power and latency.

However, still it is an ongoing topic of research what applications and network architectures are best suited for SNN. Obviously, a sparse spike generation helps to reduce power and memory consumption [8]. Several approaches on spike encoding and neuron function are investigated for the gesture recognition, trading spike rate vs. accuracy [9]. For some layers, like the input layers as bitmaps, it is even beneficial to stay with e.g. CNN layers. Resulting in a hybrid CNN-SNN model still trained end-to-end.

All this learning contributes to a benchmarking of such neuromorphic approaches [10, 11]. Nevertheless, it still remains a hard problem due to large variation of use cases, platforms and training, pointing to competitions as best comparison [12].

This work was supported by my co-workers. Pascal Gerhards, Felix Kreutz generating the radar trainings data for the gesture recognition, setup the application and building 1st models, trained using the concept of surrogate gradients. Daniel Scholz helped to further improve the model and prepare for implementation on SpiNNaker2. Jiaxin Huang implemented the closed loop application on SpiNNaker2 FPGA. SpiNNaker2 FPGA code was provided by TU-Dresden, Christian Mayr and Bernhard Vogginger.

The work was funded by the German Federal Ministry of Education and Research (BMBF) within the KI-ASIC project (16ES0993).

© FUNDACIO DE LA COMUNITAT VALENCIANA SCITO
We use our own and third party cookies for analysing and measuring usage of our website to improve our services. If you continue browsing, we consider accepting its use. You can check our Cookies Policy in which you will also find how to configure your web browser for the use of cookies. More info