JOURNAL ARTICLE

Block-wise Pruning for Convolutional Neural Networks

Abstract

As one of the mainstream model compression techniques, network pruning has received wide attention. Existing network pruning is mainly divided into weight pruning, channel pruning, and other new types. While weight pruning usually implies the highest compression ratio, it leads to high index storage and decoding overhead costs due to the irregular structure. Channel pruning, as a structured pruning method, has been widely studied. However, channel pruning tends to cause more residuals of redundant parameters. To make the pruned model achieve a better trade-off between compression ratio and regularly sparse, in this paper, we propose a novel type of network pruning named block-wise pruning (BWP). Inspired by group convolution, BWP performs pruning from a new granularity, which is beneficial for identifying and removing more residuals. Moreover, we propose a new interval-constrained penalty term, given a pruning threshold, which can make more parameters sparse, resulting in a higher compression ratio. We evaluate the effectiveness of our method on popular benchmark datasets. Compared with some state-of-the-art methods, our method shows obvious superiority. For example, with VGG-16, we achieve an 86.48% FLOPs reduction by removing 95.86% of the parameters, with only a small loss of 0.30% in accuracy on CIFAR-10. With ResNet-34, we achieve an 82.27% FLOPs reduction by removing 58.60% of the parameters, with only a loss of 1.95% in the top-1 accuracy and 1.47% in the top-5 accuracy on ImageNet.

Keywords:
Computer science Convolutional neural network Pruning Block (permutation group theory) Artificial intelligence Pattern recognition (psychology) Mathematics

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
36
Refs
0.21
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Neural Networks and Applications
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

JOURNAL ARTICLE

Pruning Ratio Optimization with Layer-Wise Pruning Method for Accelerating Convolutional Neural Networks

Koji KAMMASarimu INOUEToshikazu Wada

Journal:   IEICE Transactions on Information and Systems Year: 2021 Vol: E105.D (1)Pages: 161-169
JOURNAL ARTICLE

1xN Pattern for Pruning Convolutional Neural Networks

Mingbao LinYuxin ZhangYuchao LiBohong ChenFei ChaoMengdi WangShen LiYonghong TianRongrong Ji

Journal:   IEEE Transactions on Pattern Analysis and Machine Intelligence Year: 2022 Vol: 45 (4)Pages: 1-11
© 2026 ScienceGate Book Chapters — All rights reserved.