High-speed emerging memories for AI hardware accelerators

  • Lu A
  • Lee J
  • Kim T
  • et al.
N/ACitations
Citations of this article
41Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Applications of artificial intelligence (AI) necessitate AI hardware accelerators able to efficiently process data-intensive and computation-intensive AI workloads. AI accelerators require two types of memory: the weight memory that stores the parameters of the AI models and the buffer memory that stores the intermediate input or output data when computing a portion of the AI models. In this Review, we present the recent progress in the emerging high-speed memory for AI hardware accelerators and survey the technologies enabling the global buffer memory in digital systolic-array architectures. Beyond conventional static random-access memory (SRAM), we highlight the following device candidates: capacitorless gain cell-based embedded dynamic random-access memories (eDRAMs), ferroelectric memories, spin-transfer torque magnetic random-access memory (STT-MRAM) and spin-orbit torque magnetic random-access memory (SOT-MRAM). We then summarize the research advances in the industrial development and the technological challenges in buffer memory applications. Finally, we present a systematic benchmarking analysis on a tensor processing unit (TPU)-like AI accelerator in the edge and in the cloud and evaluate the use of these emerging memories.

Cite

CITATION STYLE

APA

Lu, A., Lee, J., Kim, T.-H., Karim, M. A. U., Park, R. S., Simka, H., & Yu, S. (2024). High-speed emerging memories for AI hardware accelerators. Nature Reviews Electrical Engineering, 1(1), 24–34. https://doi.org/10.1038/s44287-023-00002-9

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free