Abstract
The blistering progress of convolutional neural networks (CNNs) in numerous applications of the real-world usually obstruct by a surge in network volume and computational cost. Recently, researchers concentrate on eliminating these issues by compressing the CNN models, such as pruning filters and weights. In comparison with the technique of pruning weights, the technique of pruning filters doesn’t effect in sparse connectivity patterns. In this article, we have proposed a fresh new technique to estimate the significance of filters. More precisely, we combined L1-norm with capped L1-norm to represent the amount of information extracted by the filter and control regularization. In the process of pruning, the insignificant filters remove directly without any loss in the test accuracy, providing much slimmer and compact models with comparable accuracy and this process is iterated a few times. To validate the effectiveness of our algorithm. We experimentally determine the usefulness of our approach with several advanced CNN models on numerous standard data sets. Particularly, data sets CIFAR-10 is used on VGG-16 and prunes 92.7% parameters with float-point-operations (FLOPs) reduction of 75.8% without loss of accuracy and has achieved advancement in state-of-art.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 770–778
Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 580–587
Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 3431–3440
Noh H, Hong S, Han B (2015) Learning deconvolution network for semantic segmentation. In: Proceedings of the IEEE international conference on computer vision. pp 1520–1528
Wu J, Leng C, Wang Y, et al (2016) Quantized convolutional neural networks for mobile devices. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp 4820–4828
Kadav A, Durdanovic I, Graf HP (2018) Pruning filters for efficient convolutional neural networks for image recognition in surveillance applications. Google Patents
LeCun Y, Denker JS, Solla SA (1990) Optimal brain damage. In: Advances in neural information processing systems. pp 598–605
Hassibi B, Stork DG (1993) Second order derivatives for network pruning: optimal brain surgeon. In: Advances in neural information processing systems. pp 164–171
Mariet Z, Sra S (2015) Diversity networks: neural network compression using determinantal point processes. ArXiv Prepr ArXiv151105077
Han S, Mao H, Dally WJ (2015) Deep compression: compressing deep neural networks with pruning, trained quantization and huffman coding. ArXiv Prepr ArXiv151000149
Li H, Kadav A, Durdanovic I, et al (2016) Pruning filters for efficient convnets. ArXiv Prepr ArXiv160808710
He Y, Lin J, Liu Z, et al (2018) Amc: Automl for model compression and acceleration on mobile devices. In: Proceedings of the European conference on computer vision (ECCV). p 784–800
Luo J-H, Wu J, Lin W (2017) Thinet: A filter level pruning method for deep neural network compression. In: Proceedings of the IEEE international conference on computer vision. pp 5058–5066
Molchanov P, Tyree S, Karras T, et al (2016) Pruning convolutional neural networks for resource efficient inference. ArXiv Prepr ArXiv161106440
Jin X, Yuan X, Feng J, Yan S (2016) Training skinny deep neural networks with iterative hard thresholding methods. ArXiv Prepr ArXiv160705423
Li Y, Wang L, Peng S, Kumar A, Yin B (2019) Using feature entropy to guide filter pruning for efficient convolutional networks. In: Tetko IV, Kůrková V, Karpov P, Theis F (eds) Artificial neural networks and machine learning – ICANN 2019: deep learning. Springer International Publishing, Cham, pp 263–274
Hu H, Peng R, Tai Y-W, Tang C-K (2016) Network trimming: a data-driven neuron pruning approach towards efficient deep architectures. CoRR abs/1607.03250
Yu X, Liu T, Wang X, Tao D (2017) On compressing deep models by low rank and sparse decomposition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp 7370–7379
Jia K, Tao D, Gao S, Xu X (2017) Improving training of deep neural networks via singular value bounding. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp 4344–4352
Chen W, Wilson J, Tyree S, et al (2015) Compressing neural networks with the hashing trick. In: International Conference on Machine Learning. pp 2285–2294
Zhan Z, Gong Y, Li Z, et al (2020) A privacy-preserving DNN pruning and Mobile acceleration framework. ArXiv Prepr ArXiv200306513
Zoph B, Le QV (2016) Neural architecture search with reinforcement learning. ArXiv Prepr ArXiv161101578
Hüttenrauch M, Šošić A, Neumann G (2019) Deep reinforcement learning for swarm systems. J Mach Learn Res 20:1–31
de Bruin T, Kober J, Tuyls K, Babuška R (2018) Experience selection in deep reinforcement learning for control. J Mach Learn Res 19:1–56
Liu X, Zhu X, Li M et al (2019) Multiple kernel $ k $ k-means with incomplete kernels. IEEE Trans Pattern Anal Mach Intell 42:1191–1204
Yu X, Ye X, Gao Q (2020) Infrared handprint image restoration algorithm based on apoptotic mechanism. IEEE Access 8:47334–47343
You Z, Yan K, Ye J, et al (2019) Gate decorator: global filter pruning method for accelerating deep convolutional neural networks. In: Advances in Neural Information Processing Systems. pp 2133–2144
Garg I, Panda P, Roy K (2019) A low effort approach to structured cnn design using pca. IEEE Access 8:1347–1360
Lee N, Ajanthan T, Torr PH (2018) Snip: single-shot network pruning based on connection sensitivity. ArXiv Prepr ArXiv181002340
He Y, Kang G, Dong X, et al (2018) Soft filter pruning for accelerating deep convolutional neural networks. CoRR abs/1808.06866
Roy S, Panda P, Srinivasan G, Raghunathan A (2020) Pruning filters while training for efficiently optimizing deep learning networks. ArXiv Prepr ArXiv200302800
Yue L, Weibin Z, Lin S (2019) Really should we pruning after model be totally trained? Pruning based on a small amount of training. ArXiv Prepr ArXiv190108455
Scardapane S, Comminiello D, Hussain A, Uncini A (2017) Group sparse regularization for deep neural networks. Neurocomputing 241:81–89
Zhong J, Ding G, Guo Y, et al (2018) Where to prune: using LSTM to guide end-to-end pruning. In: IJCAI. pp 3205–3211
Sun X, Zhou D, Pan X, et al (2019) Pruning filters with L1-norm and standard deviation for CNN compression. In: Eleventh international conference on machine vision (ICMV 2018). International Society for Optics and Photonics, p 110412J
Xu Z, Huang G, Weinberger KQ, Zheng AX (2014) Gradient boosted feature selection. In: proceedings of the 20th ACM SIGKDD international conference on knowledge discovery and data mining. ACM, pp 522–531
Tuv E, Borisov A, Runger G, Torkkola K (2009) Feature selection with ensembles, artificial variables, and redundancy elimination. J Mach Learn Res 10:1341–1366
Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. ArXiv Prepr ArXiv14091556
Gonzalez TF (2007) Approximation algorithms for multilevel graph partitioning. In: Handbook of Approximation Algorithms and Metaheuristics. Chapman and Hall/CRC, pp 943–958
Huang G, Liu Z, Van Der Maaten L, Weinberger KQ (2017) Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 4700–4708
He K, Zhang X, Ren S, Sun J (2015) Delving deep into rectifiers: surpassing human-level performance on imagenet classification. In: Proceedings of the IEEE international conference on computer vision. pp 1026–1034
Liu Z, Li J, Shen Z, et al (2017) Learning efficient convolutional networks through network slimming. In: Proceedings of the IEEE International Conference on Computer Vision. pp 2736–2744
Aketi SA, Roy S, Raghunathan A, Roy K (2020) Gradual Channel pruning while training using feature relevance scores for convolutional neural networks. ArXiv Prepr ArXiv200209958
He Y, Zhang X, Sun J (2017) Channel pruning for accelerating very deep neural networks. In: Proceedings of the IEEE International Conference on Computer Vision. pp. 1389–1397
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Kumar, A., Shaikh, A.M., Li, Y. et al. Pruning filters with L1-norm and capped L1-norm for CNN compression. Appl Intell 51, 1152–1160 (2021). https://doi.org/10.1007/s10489-020-01894-y
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-020-01894-y