BS2: Bit-Serial Architecture Exploiting Weight Bit Sparsity for Efficient Deep Learning Acceleration

Eunseo Kim, Subean Lee, Chaeyun Kim, Ha Young Lim, Jimin Nam, Jaehyeong Sim

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Most of weights in deep learning models are small, thus they show high bit sparsity in MSBs. Based on this observation, we propose a bit-serial processing architecture (BS2) that exploits such bit sparsity to maximize computing efficiency. In this architecture, a bit feed front-end (BFFE) logic selects only effectual operands that will be fed into a bit-serial adder-tree. We also introduce an offline weight permutation algorithm to avoid conflicts in the BFFE. Evaluation results show that our design outperforms the previous bit-serial architecture in terms of performance and energy efficiency.

Original languageEnglish
Title of host publicationProceedings - International SoC Design Conference 2024, ISOCC 2024
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages356-357
Number of pages2
ISBN (Electronic)9798350377088
DOIs
StatePublished - 2024
Event21st International System-on-Chip Design Conference, ISOCC 2024 - Sapporo, Japan
Duration: 19 Aug 202422 Aug 2024

Publication series

NameProceedings - International SoC Design Conference 2024, ISOCC 2024

Conference

Conference21st International System-on-Chip Design Conference, ISOCC 2024
Country/TerritoryJapan
CitySapporo
Period19/08/2422/08/24

Bibliographical note

Publisher Copyright:
© 2024 IEEE.

Keywords

  • Bit Serial
  • Bit Sparsity
  • Deep Learning Acceleration

Fingerprint

Dive into the research topics of 'BS2: Bit-Serial Architecture Exploiting Weight Bit Sparsity for Efficient Deep Learning Acceleration'. Together they form a unique fingerprint.

Cite this