JOURNAL ARTICLE

PASS: Exploiting Post-Activation Sparsity in Streaming Architectures for CNN Acceleration

Abstract

With the ever-growing popularity of Artificial Intelligence, there is an increasing demand for more performant and efficient underlying hardware. Convolutional Neural Networks (CNN) are a workload of particular importance, which achieve high accuracy in computer vision applications. Inside CNNs, a significant number of the post-activation values are zero, resulting in many redundant computations. Recent works have explored this post-activation sparsity on instruction-based CNN accelerators but not on streaming CNN accelerators, despite the fact that streaming architectures are considered the leading design methodology in terms of performance. In this paper, we highlight the challenges associated with exploiting post-activation sparsity for performance gains in streaming CNN accelerators, and demonstrate our approach to address them. Using a set of modern CNN benchmarks, our streaming sparse accelerators achieve 1.41 x to 1.93 x efficiency (GOP/sDSP) compared to state-of-the-art instruction-based sparse accelerators.

Keywords:
Computer science Convolutional neural network Workload Set (abstract data type) Artificial intelligence Popularity Acceleration Deep learning Cellular neural network Hardware acceleration Computer engineering Artificial neural network Computer architecture Field-programmable gate array Embedded system Operating system

Metrics

2
Cited By
0.36
FWCI (Field Weighted Citation Impact)
30
Refs
0.55
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Advanced Neural Network Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Advanced Memory and Neural Computing
Physical Sciences →  Engineering →  Electrical and Electronic Engineering
Adversarial Robustness in Machine Learning
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

JOURNAL ARTICLE

Exploiting Dynamic Bit Sparsity in Activation for Deep Neural Network Acceleration

Yongshuai SunMengyu GuoDacheng LiangShan TangNaifeng Jing

Journal:   2021 IEEE 14th International Conference on ASIC (ASICON) Year: 2021 Pages: 1-4
JOURNAL ARTICLE

Accelerating SLIDE: Exploiting Sparsity on Accelerator Architectures

Sho KoAlexander RuckerYaqi ZhangPaul MureKunle Olukotun

Journal:   2022 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) Year: 2022 Pages: 663-670
JOURNAL ARTICLE

Exploiting sparsity and equation-free architectures in complex systems

Joshua L. ProctorSteven L. BruntonBingni W. BruntonJ. Nathan Kutz

Journal:   The European Physical Journal Special Topics Year: 2014 Vol: 223 (13)Pages: 2665-2684
BOOK-CHAPTER

Exploiting Sparsity

Vivienne SzeYu‐Hsin ChenTien-Ju YangJoel Emer

Synthesis lectures on computer architecture Year: 2020 Pages: 167-228
© 2026 ScienceGate Book Chapters — All rights reserved.