NAND-net: Minimizing computational complexity of in-memory processing for binary neural networks

Hyeonuk Kim, Jaehyeong Sim, Yeongjae Choi, Lee Sup Kim

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

29 Scopus citations

Abstract

Popular deep learning technologies suffer from memory bottlenecks, which significantly degrade the energyefficiency, especially in mobile environments. In-memory processing for binary neural networks (BNNs) has emerged as a promising solution to mitigate such bottlenecks, and various relevant works have been presented accordingly. However, their performances are severely limited by the overheads induced by the modification of the conventional memory architectures. To alleviate the performance degradation, we propose NANDNet, an efficient architecture to minimize the computational complexity of in-memory processing for BNNs. Based on the observation that BNNs contain many redundancies, we decomposed each convolution into sub-convolutions and eliminated the unnecessary operations. In the remaining convolution, each binary multiplication (bitwise XNOR) is replaced by a bitwise NAND operation, which can be implemented without any bit cell modifications. This NAND operation further brings an opportunity to simplify the subsequent binary accumulations (popcounts). We reduced the operation cost of those popcounts by exploiting the data patterns of the NAND outputs. Compared to the prior state-of-the-art designs, NAND-Net achieves 1.04-2.4x speedup and 34-59% energy saving, thus making it a suitable solution to implement efficient in-memory processing for BNNs.

Original languageEnglish
Title of host publicationProceedings - 25th IEEE International Symposium on High Performance Computer Architecture, HPCA 2019
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages661-673
Number of pages13
ISBN (Electronic)9781728114446
DOIs
StatePublished - 26 Mar 2019
Event25th IEEE International Symposium on High Performance Computer Architecture, HPCA 2019 - Washington, United States
Duration: 16 Feb 201920 Feb 2019

Publication series

NameProceedings - 25th IEEE International Symposium on High Performance Computer Architecture, HPCA 2019

Conference

Conference25th IEEE International Symposium on High Performance Computer Architecture, HPCA 2019
Country/TerritoryUnited States
CityWashington
Period16/02/1920/02/19

Bibliographical note

Publisher Copyright:
© 2019 IEEE.

Keywords

  • BNN
  • Convolution
  • In-memory processing

Fingerprint

Dive into the research topics of 'NAND-net: Minimizing computational complexity of in-memory processing for binary neural networks'. Together they form a unique fingerprint.

Cite this