JOURNAL ARTICLE

Learning Sparse Neural Networks Through Mixture-Distributed Regularization

Abstract

L 0 -norm regularization is one of the most efficient approaches to learn a sparse neural network. Due to its discrete nature, differentiable and approximate regularizations based on the concrete distribution [31] or its variants are proposed as alternatives; however, the concrete relaxation suffers from high-variance gradient estimates and is limited to its own concrete distribution. To address these issues, in this paper, we propose a more general framework for relaxing binary gates through mixture distributions. With the proposed method, any mixture pair of distributions converging to δ(0) and δ(1) can be applied to construct smoothed binary gates. We further introduce a reparameterization method for the smoothed binary gates drawn from mixture distributions to enable efficient gradient gradient-based optimization under the proposed deep learning algorithm. Extensive experiments are conducted, and the results show that the proposed approach achieves better performance in terms of pruned architectures, structured sparsity and the reduced number of floating point operations (FLOPs) as compared with other state-of-the-art sparsity-inducing methods.

Keywords:
Differentiable function Binary number Computer science Regularization (linguistics) Gradient descent Algorithm FLOPS Artificial neural network Deep neural networks Deep learning Artificial intelligence Mathematics Parallel computing

Metrics

6
Cited By
0.56
FWCI (Field Weighted Citation Impact)
86
Refs
0.60
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Sparse and Compressive Sensing Techniques
Physical Sciences →  Engineering →  Computational Mechanics
Image and Signal Denoising Methods
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Image Enhancement Techniques
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition

Related Documents

JOURNAL ARTICLE

Sparse Learning for Neural Networks with A Generalized Sparse Regularization

Anda TangTongsheng YaoLingfeng NiuYong Shi

Journal:   Procedia Computer Science Year: 2022 Vol: 214 Pages: 747-754
JOURNAL ARTICLE

Learning Sparse Neural Networks Using Non-Convex Regularization

Mohammad Khalid PanditRoohie NaazMohammad Ahsan Chishti

Journal:   IEEE Transactions on Emerging Topics in Computational Intelligence Year: 2021 Vol: 6 (2)Pages: 287-299
JOURNAL ARTICLE

Learning Sparse Low-Precision Neural Networks With Learnable Regularization

Yoojin ChoiMostafa El‐KhamyJungwon Lee

Journal:   IEEE Access Year: 2020 Vol: 8 Pages: 96963-96974
JOURNAL ARTICLE

Nonconvex regularization for sparse neural networks

Konstantin PieperArmenak Petrosyan

Journal:   Applied and Computational Harmonic Analysis Year: 2022 Vol: 61 Pages: 25-56
JOURNAL ARTICLE

Group sparse regularization for deep neural networks

Simone ScardapaneDanilo ComminielloAmir HussainAurelio Uncini

Journal:   Neurocomputing Year: 2017 Vol: 241 Pages: 81-89
© 2026 ScienceGate Book Chapters — All rights reserved.