Abstract
Processing-in-Memory (PIM) is an emerging computing architecture that has gained significant attention in recent times. It aims to maximize data movement efficiency by moving away from the traditional von Neumann architecture. PIM is particularly well-suited for handling deep neural networks (DNNs) that require significant data movement between the processing unit and the memory device. As a result, there has been substantial research in this area. To optimally handle DNNs with diverse structures and inductive biases, such as convolutional neural networks, graph convolutional networks, recurrent neural networks, and transformers, within a PIM architecture, careful consideration should be given to how data mapping and data flow are processed in PIM. This paper aims to provide insight into these aspects by analyzing the characteristics of various DNNs and providing detailed explanations of how they have been implemented with PIM architectures using commercially available memory technologies like DRAM and next-generation memory technologies like ReRAM.
| Original language | English |
|---|---|
| Pages (from-to) | 322-339 |
| Number of pages | 18 |
| Journal | Journal of Semiconductor Technology and Science |
| Volume | 23 |
| Issue number | 5 |
| DOIs | |
| State | Published - Oct 2023 |
Keywords
- deep learning
- deep neural network
- near-memory computing
- next-generation memory
- Processing-in-memory
Fingerprint
Dive into the research topics of 'In-depth Survey of Processing-in-memory Architectures for Deep Neural Networks'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver