In-depth Survey of Processing-in-memory Architectures for Deep Neural Networks

Ji Hoon Jang, Jin Shin, Jun Tae Park, In Seong Hwang, Hyun Kim

Research output: Contribution to journalArticlepeer-review

6 Scopus citations

Abstract

Processing-in-Memory (PIM) is an emerging computing architecture that has gained significant attention in recent times. It aims to maximize data movement efficiency by moving away from the traditional von Neumann architecture. PIM is particularly well-suited for handling deep neural networks (DNNs) that require significant data movement between the processing unit and the memory device. As a result, there has been substantial research in this area. To optimally handle DNNs with diverse structures and inductive biases, such as convolutional neural networks, graph convolutional networks, recurrent neural networks, and transformers, within a PIM architecture, careful consideration should be given to how data mapping and data flow are processed in PIM. This paper aims to provide insight into these aspects by analyzing the characteristics of various DNNs and providing detailed explanations of how they have been implemented with PIM architectures using commercially available memory technologies like DRAM and next-generation memory technologies like ReRAM.

Original languageEnglish
Pages (from-to)322-339
Number of pages18
JournalJournal of Semiconductor Technology and Science
Volume23
Issue number5
DOIs
StatePublished - Oct 2023

Keywords

  • deep learning
  • deep neural network
  • near-memory computing
  • next-generation memory
  • Processing-in-memory

Fingerprint

Dive into the research topics of 'In-depth Survey of Processing-in-memory Architectures for Deep Neural Networks'. Together they form a unique fingerprint.

Cite this