Abstract

hough ReRAM-based convolutional neural network (CNN) accelerators have been widely studied, state-of-the-art solutions suffer from either incapability of training (e.g., ISSAC [1]) or inefficiency of inference (e.g., PipeLayer [2]) due to the pipeline design. In this work, we propose AtomLayer---a universal ReRAM-based accelerator to support both efficient CNN training and inference. AtomLayer uses the atomic layer computation which processes only one network layer each time to eliminate the pipeline related issues such as long latency, pipeline bubbles and large on-chip buffer overhead. For further optimization, we use a unique filter mapping and a data reuse system to minimize the cost of layer switching and DRAM access. Our experimental results show that AtomLayer can achieve higher power efficiency than ISSAC in inference (1.1×) and PipeLayer in training (1.6×), respectively, meanwhile reducing the footprint by 15×.


Original document

The different versions of the original document can be found in:

https://ieeexplore.ieee.org/document/8465832,
https://doi.acm.org/10.1145/3195970.3195998,
https://doi.org/10.1145/3195970.3195998,
https://dl.acm.org/citation.cfm?id=3195998,
https://academic.microsoft.com/#/detail/2809171749
http://dx.doi.org/10.1145/3195970.3195998 under the license http://www.acm.org/publications/policies/copyright_policy#Background
Back to Top

Document information

Published on 01/01/2018

Volume 2018, 2018
DOI: 10.1145/3195970.3195998
Licence: Other

Document Score

0

Views 0
Recommendations 0

Share this document

Keywords

claim authorship

Are you one of the authors of this document?