Login / Signup

Precision-extension technique for accurate vector-matrix multiplication with a CNT transistor crossbar array.

Sungho KimYongwoo LeeHee-Dong KimSung-Jin Choi
Published in: Nanoscale (2019)
Most machine learning algorithms involve many multiply-accumulate operations, which dictate the computation time and energy required. Vector-matrix multiplications can be accelerated using resistive networks, which can be naturally implemented in a crossbar geometry by leveraging Kirchhoff's current law in a single readout step. However, practical computing tasks that require high precision are still very challenging to implement in a resistive crossbar array owing to intrinsic device variability and unavoidable crosstalk, such as sneak path currents through adjacent devices, which inherently result in low precision. Here, we experimentally demonstrate a precision-extension technique for a carbon nanotube (CNT) transistor crossbar array. High precision is attained through multiple devices operating together, each of which stores a portion of the required bit width. A 10 × 10 CNT transistor array can perform vector-matrix multiplication with high accuracy, making in-memory computing approaches attractive for high-performance computing environments.
Keyphrases
  • machine learning
  • high resolution
  • high throughput
  • carbon nanotubes
  • working memory
  • high density
  • artificial intelligence
  • deep learning
  • big data
  • single cell