Neuromorphic Computing and Engineering (Jan 2023)

System model of neuromorphic sequence learning on a memristive crossbar array

  • Sebastian Siegel,
  • Younes Bouhadjar,
  • Tom Tetzlaff,
  • Rainer Waser,
  • Regina Dittmann,
  • Dirk J Wouters

DOI
https://doi.org/10.1088/2634-4386/acca45
Journal volume & issue
Vol. 3, no. 2
p. 024002

Abstract

Read online

Machine learning models for sequence learning and processing often suffer from high energy consumption and require large amounts of training data. The brain presents more efficient solutions to how these types of tasks can be solved. While this has inspired the conception of novel brain-inspired algorithms, their realizations remain constrained to conventional von-Neumann machines. Therefore, the potential power efficiency of the algorithm cannot be exploited due to the inherent memory bottleneck of the computing architecture. Therefore, we present in this paper a dedicated hardware implementation of a biologically plausible version of the Temporal Memory component of the Hierarchical Temporal Memory concept. Our implementation is built on a memristive crossbar array and is the result of a hardware-algorithm co-design process. Rather than using the memristive devices solely for data storage, our approach leverages their specific switching dynamics to propose a formulation of the peripheral circuitry, resulting in a more efficient design. By combining a brain-like algorithm with emerging non-volatile memristive device technology we strive for maximum energy efficiency. We present simulation results on the training of complex high-order sequences and discuss how the system is able to predict in a context-dependent manner. Finally, we investigate the energy consumption during the training and conclude with a discussion of scaling prospects.

Keywords