Skip to content

MrLinNing/MemristorLSM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

32 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Resistive memory-based zero-shot liquid state machine for multimodal event data learning

The human brain is a complex spiking neural network (SNN) that learns multimodal signals in a zero-shot manner by generalizing existing knowledge. Remarkably, the brain achieves this with minimal power consumption, using event-based signals that propagate within its structure.
However, mimicking the human brain in neuromorphic hardware presents both hardware and software challenges. Hardware limitations, such as the slowdown of Moore's law and the von Neumann bottleneck, hinder the efficiency of digital computers. On the software side, SNNs are known for their difficult training, especially when learning multimodal signals.

To overcome these challenges, we propose a hardware-software co-design that combines a fixed and random liquid state machine (LSM) SNN encoder with trainable artificial neural network (ANN) projections. The LSM is physically implemented using analogue resistive memory, leveraging the inherent stochasticity of resistive switching to generate random weights. This highly efficient and nanoscale in-memory computing approach effectively addresses the von Neumann bottleneck and the slowdown of Moore's law. The ANN projections are implemented digitally, allowing for easy optimization using contrastive loss, which helps to overcome the difficulties associated with SNN training.

We experimentally implement this co-design on a 40 nm 256 Kb in-memory computing macro. We first demonstrate LSM-based event encoding through supervised classification and linear probing on the N-MNIST and N-TIDIGITS datasets. Based on that, we showcase the zero-shot learning of multimodal events, including visual and audio data association, as well as neural and visual data alignment for brain-machine interfaces. Our co-design achieves classification accuracy comparable to fully optimized software models. This not only results in a 152.83 and 393.07-fold reduction in training costs compared to state-of-the-art contrastive language-image pre-training (CLIP) and Prototypical networks, but also delivers a 23.34 and 161-fold improvement in energy efficiency compared to cutting-edge digital hardware, respectively.

These proof-of-principle prototypes not only demonstrate the efficient and compact neuromorphic hardware using in-memory computing, but also zero-shot learning multimodal events in a brain-inspired manner, paving the way for future edge neuromorphic intelligence.

framework

Dataset

For event-driven N-MNIST and N-TIDIGITS datasets, please download from One drive.

EEG Dataset can be download from Willett, Francis et al. (2021), Data from: High-performance brain-to-text communication via handwriting, Dryad, Dataset, https://doi.org/10.5061/dryad.wh70rxwmv

For the E-MNIST, the data will be downloaded from the torch dataloader.

The EEG and E-MNIST datasets have been processed to be event-driven data by data CLIP and RateEncoding methods, respectively. Please see the data_process_enlarge.py file in the EEG_clip/EEG_to_Image folder for details.

Hardware and Software Information

Hardware

GPU: NVIDIA GeForce RTX 3090 Ti
CPU: AMD Ryzen 9 7950X 16-Core Processor
Memory: 61G

Software

Python version:  3.9.12 (main, Apr  5 2022, 06:56:58) 
[GCC 7.5.0]
torch version:  2.0.1
numpy version:  1.21.5
matplotlib version:  3.5.1
sklearn version:  1.0.2
CUDA is available
CUDA version:  11.7

LSM-based Supervised Classification on N-MNIST

cd LSM-NMNIST

LSM-based Supervised Classification on N-TIDIGITS

cd LSM-NTIDIGITS

LSM-based Contrastive Learning for N-MNIST and N-TIDIGITS

cd LSM-CLIP-ATV

LSM-based Contrastive Learning for EEG and E-MNIST

cd LSM-CLIP-BTV

LICENSE

MIT License

Copyright (c) Department of Electrical and Electronic Engineering, the University of Hong Kong All rights reserved.

Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:

The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.

THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published