TY - JOUR
T1 - Resistive Neural Hardware Accelerators
AU - Smagulova, Kamilya
AU - Fouda, Mohamed E.
AU - Kurdahi, Fadi
AU - Salama, Khaled N.
AU - Eltawil, Ahmed
N1 - KAUST Repository Item: Exported on 2023-05-19
Acknowledged KAUST grant number(s): URF/1/4704-01-01
Acknowledgements: This work was supported by the King Abdullah University of Science and Technology through the Competitive Research Grant (CRG) under Grant URF/1/4704-01-01.
PY - 2023/5/16
Y1 - 2023/5/16
N2 - Deep neural networks (DNNs), as a subset of machine learning (ML) techniques, entail that real-world data can be learned, and decisions can be made in real time. However, their wide adoption is hindered by a number of software and hardware limitations. The existing general-purpose hardware platforms used to accelerate DNNs are facing new challenges associated with the growing amount of data and are exponentially increasing the complexity of computations. Emerging nonvolatile memory (NVM) devices and the compute-in-memory (CIM) paradigm are creating a new hardware architecture generation with increased computing and storage capabilities. In particular, the shift toward resistive random access memory (ReRAM)-based in-memory computing has great potential in the implementation of area- and power-efficient inference and in training large-scale neural network architectures. These can accelerate the process of IoT-enabled AI technologies entering our daily lives. In this survey, we review the state-of-the-art ReRAM-based DNN many-core accelerators, and their superiority compared to CMOS counterparts was shown. The review covers different aspects of hardware and software realization of DNN accelerators, their present limitations, and prospects. In particular, a comparison of the accelerators shows the need for the introduction of new performance metrics and benchmarking standards. In addition, the major concerns regarding the efficient design of accelerators include a lack of accuracy in simulation tools for software and hardware codesign.
AB - Deep neural networks (DNNs), as a subset of machine learning (ML) techniques, entail that real-world data can be learned, and decisions can be made in real time. However, their wide adoption is hindered by a number of software and hardware limitations. The existing general-purpose hardware platforms used to accelerate DNNs are facing new challenges associated with the growing amount of data and are exponentially increasing the complexity of computations. Emerging nonvolatile memory (NVM) devices and the compute-in-memory (CIM) paradigm are creating a new hardware architecture generation with increased computing and storage capabilities. In particular, the shift toward resistive random access memory (ReRAM)-based in-memory computing has great potential in the implementation of area- and power-efficient inference and in training large-scale neural network architectures. These can accelerate the process of IoT-enabled AI technologies entering our daily lives. In this survey, we review the state-of-the-art ReRAM-based DNN many-core accelerators, and their superiority compared to CMOS counterparts was shown. The review covers different aspects of hardware and software realization of DNN accelerators, their present limitations, and prospects. In particular, a comparison of the accelerators shows the need for the introduction of new performance metrics and benchmarking standards. In addition, the major concerns regarding the efficient design of accelerators include a lack of accuracy in simulation tools for software and hardware codesign.
UR - http://hdl.handle.net/10754/671210
UR - https://ieeexplore.ieee.org/document/10127587/
U2 - 10.1109/jproc.2023.3268092
DO - 10.1109/jproc.2023.3268092
M3 - Article
SN - 0018-9219
VL - 111
SP - 500
EP - 527
JO - Proceedings of the IEEE
JF - Proceedings of the IEEE
IS - 5
ER -