JOURNAL ARTICLE

An Efficient Accelerator for Sparse Convolutional Neural Networks

Abstract

In this paper, we propose a sparse convolutional neural network accelerator design on FPGAs. Similar to the DNNWEAVER architecture, our accelerator uses two-level hierarchy: multiple Processing Units (PUs) and each PU comprises a set of Processing Elements (PEs). To address the irregularity of sparse neural networks, we introduce a novel sparse dataflow for sparse CNN computing as well as weight merging method to balance the computation load on different PUs for better overall efficiency. We implement our design with 32 PUs and 14 PEs in each PU. When compared with DNNWEAVER on VGG16 network, our accelerator achieves 3.49× speedup and 3.05× energy saving on average when running at 150MHz on a Xilinx ZC706 board and reaches the speed of 400 GOPS.

Keywords:
Convolutional neural network Computer science Convolutional code Artificial intelligence Algorithm Decoding methods

Metrics

4
Cited By
0.32
FWCI (Field Weighted Citation Impact)
14
Refs
0.62
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Advanced Neural Network Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Advanced Memory and Neural Computing
Physical Sciences →  Engineering →  Electrical and Electronic Engineering
Machine Learning and ELM
Physical Sciences →  Computer Science →  Artificial Intelligence
© 2026 ScienceGate Book Chapters — All rights reserved.