Abstract:
A class of brain-inspired recurrent algorithms known as <;i>reservoir computing (RC) networks<;/i> reduces the computational complexity and cost of training machine-learn...Show MoreMetadata
Abstract:
A class of brain-inspired recurrent algorithms known as <;i>reservoir computing (RC) networks<;/i> reduces the computational complexity and cost of training machine-learning models by using random, fixed synaptic strengths. This article offers insights about a spiking reservoir network, the liquid state machine (LSM), the inner workings of the algorithm, the design metrics, and neuromorphic designs. The discussion extends to variations of the LSM that incorporate local plasticity mechanisms and hierarchy to improve performance and memory capacity.
Published in: IEEE Signal Processing Magazine ( Volume: 36, Issue: 6, November 2019)