JOURNAL ARTICLE

Learning Sparse Neural Networks Using Non-Convex Regularization

Mohammad Khalid PanditRoohie NaazMohammad Ahsan Chishti

Year: 2021 Journal:   IEEE Transactions on Emerging Topics in Computational Intelligence Vol: 6 (2)Pages: 287-299   Publisher: Institute of Electrical and Electronics Engineers

Abstract

Deep Neural Networks (DNNs) is the computing paradigm that has achieved remarkable success in various fields of engineering in recent years, primarily visual recognition. DNNs owe its success to the presence of large number of weight parameters (and increased depth), which led to huge computation and memory costs for implementation. These limiting factors hinder the scalability of such algorithms on resource constraint devices (like IoT devices). In general, DNNs are believed to be over parametrized i.e., the parameters are highly redundant, thus can be structurally removed without significant loss of performance. To solve these issues, we propose to use non-convex T $\ell _{1}$ regularizer along with the additional effect of sparse group lasso to completely remove the redundant neurons/filters that is, to introduce structured sparsity. The network has been trained using the proximal gradient method, which is useful in optimizing functions with the combination of smooth and non-smooth terms. We show that proposed regularizer manages to achieve competitive performances as well as extremely compact networks. Detailed experiments are performed on several benchmark datasets that illustrate the efficiency of the approach. On the ImageNet dataset, our approach removes more than 50% of parameters of convolutional layers and 85% parameters of fully connected layers of Alexnet with no drop in accuracy.

Keywords:
Scalability Computer science Computation Regularization (linguistics) Benchmark (surveying) Regular polygon Deep neural networks Artificial neural network Limiting Redundancy (engineering) Convolutional neural network Algorithm Artificial intelligence Mathematics

Metrics

13
Cited By
1.76
FWCI (Field Weighted Citation Impact)
82
Refs
0.80
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Sparse and Compressive Sensing Techniques
Physical Sciences →  Engineering →  Computational Mechanics
Machine Learning and ELM
Physical Sciences →  Computer Science →  Artificial Intelligence
Domain Adaptation and Few-Shot Learning
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

© 2026 ScienceGate Book Chapters — All rights reserved.