Spiking Memristive Architectures for Learning to Learn
Spiking Memristive Architectures for Learning to Learn
Disciplines
Electrical Engineering, Electronics, Information Engineering (25%); Computer Sciences (50%); Nanotechnology (25%)
Keywords
-
Spiking Neural Networks,
Memristor,
Neuromorphic,
Analog Computing
Contemporary artificial intelligence (AI) and machine learning applications often imply heavy computational loads with current technology. However, there is a growing demand for low-power autonomously learning AI systems that are employed in the field. This scenario includes applications in mobile devices, autonomous mobile robots and vehicles, edge devices, and the internet of things, to name just a few obvious examples. The SMALL project will investigate low-power solutions for this alternative application scenario. Inspired by the architecture of the human brain, the system will be based on networks of so- called spiking neurons, neurons that communicate by sending short voltage pulses via synaptic connections. Today, neural networks are implemented in digital hardware. In contrast, we are developing hardware that implements the operation of spiking neurons using analog voltages and the specific properties of the computational substrate. This design principle naturally leads to low-power systems. Another essential innovation of the system is its memory-technolgy, which will not consist of standard memory elements, but rather of memristive memory. In contrast to usual memory technology, this novel type of memory does not only store bits but analog values, with the promise for extremely high memory density and further improvements of memory efficacy. In the field applications often demand online adaptation of such systems, which usually necessitates slow and complex training procedures. To overcome this problem, we will investigate the applicability of learning to learn (L2L) to spiking memristive hardware. In other words, the system will be trained in the laboratory to quickly learn a task in the envisioned application domain. In the application itself, the system then quickly adapts to the task at hand. In summary, the goal of the SMALL project is to build versatile and adaptive low-power small size AI machinery based on spiking neural networks with memristive synapses using L2L. As a proof of concept, we will deliver an experimental system in a real-world robotics environment.
Contemporary Artificial Intelligence (AI) applications often rely on methods that imply heavy computational loads. However, there is a growing demand for low-power autonomously learning AI systems that are employed "in the field". We investigated in this project options for learning in low-power unconventional hardware that is based on so-called spiking neural networks (SNNs). SNNs are a biologically inspired neural network type with energy-efficient spike-based communication between neurons. Our networks were implemented in so-called neuromorphic hardware, which is hardware that implements essential ingredients of biological neural networks. As synaptic connections between neurons, we considered nano-scale electric circuit elements, so-called memristors. Memristors are can be used to implement large synaptic arrays very efficiently. However, their behaviour is also noisy, which necessitates novel training schemes. "In the field" applications often demand online adaptation of such systems, which typically implies hardware-averse training procedures. To overcome this problem, we investigated the applicability of "learning to learn" (L2L) for the neuromorphic hardware. In an initial optimization, the hardware is trained to become a good learner for the target application. Here, arbitrarily complex learning algorithms can be used on a host system with the hardware "in the loop". In the application itself, simpler algorithms - that can be easily implemented in neuromorphic hardware - provide adaptation of the network in hardware. In this project, we developed novel variants of spiking neural networks that are significantly more powerful and versatile than previous models. These networks are able to learn various demanding tasks such as the evaluation of mathematical expressions. In particular, we considered spiking neural networks with brain-like memory systems which enables them for example to learn from single examples and answer questions about previously presented stories. We also developed novel learning algorithms that enable the training of neuromorphic hardware with low-precision synaptic weights as well as hardware with highly unreliable memristive synapses. Furthermore, we established paradigms to utilize learning-to-learn for neuromorphic systems. The final outcome of the project was an experimental system in a real-world robotics environment. An industrial robot arm was controlled by a spiking neural network with memristive synapses. Using the learning-to-learn approach, the robot was able to learn a novel arm trajectory after a single exposure to a demonstrated arm movement. This provides a proof of concept that learning-to-learn can be applied to neuromrophic systems with memristive synapses.
- Technische Universität Graz - 100%
- Alejandro Barranco-Linares, University of Sevilla - Spain
- Evangelos Eleftheriou, IBM Research Division, Zürich - Switzerland
- Giacomo Indiveri, University of Zurich - Switzerland
- Themis Prodromakis, University of Edinburgh
Research Output
- 81 Citations
- 12 Publications
- 4 Datasets & models
- 5 Disseminations
- 5 Scientific Awards
- 1 Fundings
-
2024
Title Fast learning without synaptic plasticity in spiking neural networks. DOI 10.1038/s41598-024-55769-0 Type Journal Article Author Bellec G Journal Scientific reports Pages 8557 -
2023
Title Quantized rewiring: hardware-aware training of sparse deep neural networks DOI 10.1088/2634-4386/accd8f Type Journal Article Author Legenstein R Journal Neuromorphic Computing and Engineering -
2025
Title Memory-Dependent Computation and Learning in Spiking Neural Networks Through Hebbian Plasticity. DOI 10.1109/tnnls.2023.3341446 Type Journal Article Author Limbacher T Journal IEEE transactions on neural networks and learning systems Pages 2551-2562 -
2023
Title Spike-based models for cognitive computations and robuts training of memristive neural networks Type PhD Thesis Author Caca Kraisnikovic -
2023
Title Utilizing synaptic plasticity and memory for computation and learning in simple neural network models of brain function Type PhD Thesis Author Thomas Limbacher -
2023
Title Fault Pruning: Robust Training ofNeural Networks withMemristive Weights; In: Unconventional Computation and Natural Computation - 20th International Conference, UCNC 2023, Jacksonville, FL, USA, March 13-17, 2023, Proceedings DOI 10.1007/978-3-031-34034-5_9 Type Book Chapter Publisher Springer Nature Switzerland -
2021
Title Spike frequency adaptation supports network computations on temporally dispersed information DOI 10.7554/elife.65459 Type Journal Article Author Salaj D Journal eLife Link Publication -
2021
Title Spike-based symbolic computations on bit strings and numbers DOI 10.1101/2021.07.14.452347 Type Preprint Author Kraišnikovic C Pages 2021.07.14.452347 Link Publication -
2020
Title H-Mem: Harnessing synaptic plasticity with Hebbian Memory Networks DOI 10.1101/2020.07.01.180372 Type Preprint Author Limbacher T Pages 2020.07.01.180372 Link Publication -
2022
Title Memory-enriched computation and learning in spiking neural networks through Hebbian plasticity DOI 10.48550/arxiv.2205.11276 Type Preprint Author Limbacher T -
2021
Title Revisiting the role of synaptic plasticity and network dynamics for fast learning in spiking neural networks DOI 10.1101/2021.01.25.428153 Type Preprint Author Subramoney A Pages 2021.01.25.428153 Link Publication -
2021
Title Many-Joint Robot Arm Control with Recurrent Spiking Neural Networks DOI 10.1109/iros51168.2021.9636001 Type Conference Proceeding Abstract Author Traub M Pages 4918-4925 Link Publication
-
2023
Link
Title Fault-Pruning Type Computer model/algorithm Public Access Link Link -
2023
Link
Title Q-Rewiring Type Computer model/algorithm Public Access Link Link -
2021
Link
Title Spiking Neural Networks with Spike-frequency adaptation Type Computer model/algorithm Public Access Link Link -
2020
Link
Title Hebbian Memory Networks Type Computer model/algorithm Public Access Link Link
-
2021
Link
Title Showing Styria Type Participation in an activity, workshop or similar Link Link -
2022
Title ASAI workshop Type A formal working group, expert panel or dialogue -
2023
Title AI Summer School lecture Type A talk or presentation -
2022
Title Telluride discussion Type A talk or presentation -
2020
Link
Title Planet Research Article Type A press release, press conference or response to a media enquiry/interview Link Link
-
2022
Title Telluride Workshop Type Personally asked as a key note speaker to a conference Level of Recognition Continental/International -
2022
Title Keynote BI 2022 Type Personally asked as a key note speaker to a conference Level of Recognition Continental/International -
2021
Title Invited talk Beyond von Neumann Architectures Workshop Type Personally asked as a key note speaker to a conference Level of Recognition Continental/International -
2020
Title Spotlight NeurIPS Type Personally asked as a key note speaker to a conference Level of Recognition Continental/International -
2020
Title Keynote Computational Neuroscience Type Personally asked as a key note speaker to a conference Level of Recognition Continental/International
-
2023
Title EFRI BRAID: Scalable-Learning Neuromorphics Type Research grant (including intramural programme) Start of Funding 2023 Funder National Science Foundation (NSF)