A Lazy Engine for High-utilization and Energy-efficient ReRAM-based Neural Network Accelerator

Resistive random-access memory (ReRAM) has been explored to be a promising solution to accelerate the inference of deep neural networks at the embedded systems by performing computations in memory. To reduce the latency of the neural network, all the pre-trained weights are pre-programmed in ReRAM c...

Full description

Saved in:
Bibliographic Details
Published in:2022 IEEE 20th International Conference on Industrial Informatics (INDIN) pp. 140 - 145
Main Authors: Yang, Wei-Yi, Chen, Ya-Shu, Xiao, Jin-Wen
Format: Conference Proceeding
Language:English
Published: IEEE 25-07-2022
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Resistive random-access memory (ReRAM) has been explored to be a promising solution to accelerate the inference of deep neural networks at the embedded systems by performing computations in memory. To reduce the latency of the neural network, all the pre-trained weights are pre-programmed in ReRAM cells as device resistance for the inference phase. However, the system utilization is decreased by the data dependency of the deployed neural networks and results in low energy efficiency. In this work, we propose a Lazy Engine for providing high utilization and energy-efficient ReRAM-based accelerators. Instead of avoiding idle time by applying ReRAM crossbar duplication, Lazy Engine delays the start time of the vector-matrix multiplication operations, with run-time programming overhead consideration, to reclaim idle time for energy efficiency while improving resource utilization. The experimental results show that Lazy Engine achieves up to 77% and 96% improvement in resource utilization and energy saving compared to state-of-the-art ReRAM-based accelerators.
DOI:10.1109/INDIN51773.2022.9976171