Overcoming phase-change material non-idealities by meta-learning for adaptation on the edge
Yigit Demirag a, Regina Dittmann b, Giacomo Indiveri a, Emre Neftci b
a The Institute of Neuroinformatics, University of Zurich and ETH Zurich, Switzerland
b Peter Grunberg Institute, Forschungszentrum Juelich, Germany
Proceedings of Neuromorphic Materials, Devices, Circuits and Systems (NeuMatDeCaS)
VALÈNCIA, Spain, 2023 January 23rd - 25th
Organizers: Rohit Abraham John, Irem Boybat, Jason Eshraghian and Simone Fabiano
Contributed talk, Yigit Demirag, presentation 043
DOI: https://doi.org/10.29363/nanoge.neumatdecas.2023.043
Publication date: 9th January 2023

Memristive crossbar arrays are built using scalable in-memory computing technologies. As such, these arrays can solve the von-Neumann bottleneck, which is the dominant factor of the energy consumption of neural network computations. They enable the deployment of large network models on compact embedded, mobile, edge devices for ultra-low power real-time inference applications [1].

However, in many scenarios, the post-deployment conditions might require a further adaptation in the network weights to compensate for domain data shifts due to sensory or environmental noise, gradual device degradation or simply fine-tuning to a new task that has not been seen during the offline training. The analog hardware non-idealities such as programming variability and non-linearity of conductance updates, the high power consumption of programming pulses, and the limited device endurance hinders the possibility of having power-efficient on-chip edge adaptation. This becomes further complicated with the data and memory- intensive requirements of gradient-based learning, such as training in batches and the lack of online and local credit assignment solutions. 

In this work, we demonstrate that bi-level optimization offers an offline training solution to overcome the impact of such non-idealities for online learning on edge. We use Model-Agnostic Meta-Learning (MAML) [3] to pre-train a feedforward spiking neural network with accurate memristive synapse models to perform few-shot learning on standard regression benchmarks. Through the differentiable stochastic device model, the outer optimization loop learns a good initialization of the memristive conductances for the deployement that the negative impact of device programming during inner loop updates on the edge is suppressed. To achieve this, we use an empirical model of phase-change material (PCM) [2], which comprises all major forms of analog hardware non-idealities based on the measurements from 10,000 devices. The model includes the programming stochasticity and non-linearity as a function of device programming history, 1/f read noise, and temporal conductance drift. We improved the offline outer loop optimization performance by carefully selecting twice- differentiable surrogate spiking function [4] and additional hardware-aware training methods, such as noise-injection with straight-through estimators [5] and update thresholding [6]. 

Our results showed that a network trained with bi-level optimization can be finetuned to adapt new tasks on the edge by applying single blind programming pulses to only the output layer memristors of the neural network. We report competitive accuracies thanks to learning to mitigate write noise, read noise and drift of phase-change memory devices.

We acknowledge the funding from the European Union’s H2020 research and innovation programme under the Marie Skłodowska-Curie grant agreement No 861153.

© FUNDACIO DE LA COMUNITAT VALENCIANA SCITO
We use our own and third party cookies for analysing and measuring usage of our website to improve our services. If you continue browsing, we consider accepting its use. You can check our Cookies Policy in which you will also find how to configure your web browser for the use of cookies. More info