{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,12]],"date-time":"2026-02-12T16:18:17Z","timestamp":1770913097484,"version":"3.50.1"},"reference-count":55,"publisher":"Springer Science and Business Media LLC","issue":"9","license":[{"start":{"date-parts":[[2022,1,14]],"date-time":"2022-01-14T00:00:00Z","timestamp":1642118400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,14]],"date-time":"2022-01-14T00:00:00Z","timestamp":1642118400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62071303"],"award-info":[{"award-number":["62071303"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61871269"],"award-info":[{"award-number":["61871269"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100021171","name":"Guangdong Basic and Applied Basic Research Foundation","doi-asserted-by":"crossref","award":["2019A1515011861"],"award-info":[{"award-number":["2019A1515011861"]}],"id":[{"id":"10.13039\/501100021171","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Shenzhen Science and Technology Projection","award":["JCYJ20190808151615540"],"award-info":[{"award-number":["JCYJ20190808151615540"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2022,7]]},"DOI":"10.1007\/s10489-021-02932-z","type":"journal-article","created":{"date-parts":[[2022,1,14]],"date-time":"2022-01-14T00:03:28Z","timestamp":1642118608000},"page":"10519-10540","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["Progressive kernel pruning CNN compression method with an adjustable input channel"],"prefix":"10.1007","volume":"52","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9666-0772","authenticated-orcid":false,"given":"Jihong","family":"Zhu","sequence":"first","affiliation":[]},{"given":"Jihong","family":"Pei","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,1,14]]},"reference":[{"key":"2932_CR1","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556"},{"key":"2932_CR2","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"2932_CR3","doi-asserted-by":"crossref","unstructured":"Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu C-Y, Berg AC (2016) SSD: Single shot multibox detector. In: European conference on computer vision, pp 21\u201337. Springer","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"2932_CR4","doi-asserted-by":"crossref","unstructured":"Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: Unified, real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 779\u2013788","DOI":"10.1109\/CVPR.2016.91"},{"key":"2932_CR5","doi-asserted-by":"crossref","unstructured":"Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3431\u20133440","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"2932_CR6","doi-asserted-by":"crossref","unstructured":"Li Y, Qi H, Dai J, Ji X, Wei Y (2017) Fully convolutional instance-aware semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 2359\u20132367","DOI":"10.1109\/CVPR.2017.472"},{"key":"2932_CR7","doi-asserted-by":"crossref","unstructured":"Yan S, Teng Y, Smith JS, Zhang B (2016) Driver behavior recognition based on deep convolutional neural networks. In: 2016 12th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery (ICNC-FSKD), IEEE, pp 636\u2013641","DOI":"10.1109\/FSKD.2016.7603248"},{"key":"2932_CR8","doi-asserted-by":"crossref","unstructured":"Rei\u00df S, Roitberg A, Haurilet M, Stiefelhagen R (2020) Deep classification-driven domain adaptation for cross-modal driver behavior recognition. In: 2020 IEEE Intelligent Vehicles Symposium (IV), IEEE, pp 1042\u20131047","DOI":"10.1109\/IV47402.2020.9304782"},{"issue":"3","key":"2932_CR9","doi-asserted-by":"publisher","first-page":"513","DOI":"10.1109\/TMM.2017.2751969","volume":"20","author":"N Takahashi","year":"2017","unstructured":"Takahashi N, Gygli M, Van Gool L (2017) Aenet: Learning deep audio features for video analysis. IEEE Transactions on Multimedia 20(3):513\u2013524","journal-title":"IEEE Transactions on Multimedia"},{"key":"2932_CR10","doi-asserted-by":"crossref","unstructured":"Chen J, Li K, Deng Q, Li K, Philip SY (2019) Distributed deep learning model for intelligent video surveillance systems with edge computing. IEEE Transactions on Industrial Informatics","DOI":"10.1109\/TII.2019.2909473"},{"key":"2932_CR11","unstructured":"Mozaffari S, Al-Jarrah OY, Dianati M, Jennings P, Mouzakitis A (2019) Deep learning-based vehicle behaviour prediction for autonomous driving applications: A review. arXiv preprint arXiv:1912.11676"},{"key":"2932_CR12","doi-asserted-by":"crossref","unstructured":"Fayjie AR, Hossain S, Oualid D, Lee D-J (2018) Driverless car: Autonomous driving using deep reinforcement learning in urban environment. In: 2018 15th International Conference on Ubiquitous Robots (UR), IEEE, pp 896\u2013901","DOI":"10.1109\/URAI.2018.8441797"},{"key":"2932_CR13","unstructured":"LeCun Y, Denker JS, Solla A (1990) Optimal brain damage. In: Advances in neural information processing systems, pp 598\u2013605"},{"key":"2932_CR14","unstructured":"Hassibi B, Stork DG (1993) Second order derivatives for network pruning: Optimal brain surgeon. In: Advances in neural information processing systems, pp 164\u2013171"},{"key":"2932_CR15","unstructured":"Han S, Pool J, Tran J, Dally W (2015) Learning both weights and connections for efficient neural network. In: Advances in neural information processing systems, pp 1135\u20131143"},{"key":"2932_CR16","unstructured":"Han S, Mao H, Dally WJ (2015) Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149"},{"key":"2932_CR17","unstructured":"Ullrich K, Meeds E, Welling M (2017) Soft weight-sharing for neural network compression. arXiv preprint arXiv:1702.04008"},{"key":"2932_CR18","unstructured":"Xiao X, Wang Z, Rajasekaran S (2019) Autoprune: Automatic network pruning by regularizing auxiliary parameters. In: Advances in Neural Information Processing Systems, pp 13681\u201313691"},{"key":"2932_CR19","unstructured":"Wang Y, Xu C, You S, Tao D, Xu C (2016) Cnnpack: Packing convolutional neural networks in the frequency domain. In: Advances in neural information processing systems, pp 253\u2013261"},{"key":"2932_CR20","unstructured":"Liu Z, Xu J, Peng X, Xiong R (2018) Frequency-domain dynamic pruning for convolutional neural networks. In: Advances in Neural Information Processing Systems, pp 1043\u20131053"},{"key":"2932_CR21","unstructured":"Li H, Kadav A, Durdanovic I, Samet H, Graf HP (2016) Pruning filters for efficient convnets. arXiv preprint arXiv:1608.08710"},{"key":"2932_CR22","doi-asserted-by":"crossref","unstructured":"He Y, Kang G, Dong X, Fu Y, Yang Y (2018) Soft filter pruning for accelerating deep convolutional neural networks. arXiv preprint arXiv:1808.06866","DOI":"10.24963\/ijcai.2018\/309"},{"key":"2932_CR23","doi-asserted-by":"crossref","unstructured":"He Y, Liu P, Wang Z, Hu Z, Yang Y (2019) Filter pruning via geometric median for deep convolutional neural networks acceleration. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 4340\u20134349","DOI":"10.1109\/CVPR.2019.00447"},{"key":"2932_CR24","doi-asserted-by":"crossref","unstructured":"Luo J-H, Wu J, Lin W (2017) ThiNet: A filter level pruning method for deep neural network compression. In: Proceedings of the IEEE international conference on computer vision, pp 5058\u20135066","DOI":"10.1109\/ICCV.2017.541"},{"key":"2932_CR25","doi-asserted-by":"crossref","unstructured":"Yu R, Li A, Chen C-F, Lai J-H, Morariu VI, Han X, Gao M, Lin C-Y, Davis LS (2018) Nisp: Pruning networks using neuron importance score propagation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 9194\u20139203","DOI":"10.1109\/CVPR.2018.00958"},{"key":"2932_CR26","unstructured":"Zhuang Z, Tan M, Zhuang B, Liu J, Guo Y, Wu Q, Huang J, Zhu J (2018) Discrimination-aware channel pruning for deep neural networks. In: Advances in Neural Information Processing Systems, pp 875\u2013886"},{"key":"2932_CR27","doi-asserted-by":"crossref","unstructured":"He Y, Zhang X, Sun J (2017) Channel pruning for accelerating very deep neural networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp 1389\u20131397","DOI":"10.1109\/ICCV.2017.155"},{"key":"2932_CR28","doi-asserted-by":"crossref","unstructured":"Liu Z, Li J, Shen Z, Huang G, Yan S, Zhang C (2017) Learning efficient convolutional networks through network slimming. In: Proceedings of the IEEE International Conference on Computer Vision, pp 2736\u20132744","DOI":"10.1109\/ICCV.2017.298"},{"key":"2932_CR29","doi-asserted-by":"crossref","unstructured":"Lin S, Ji R, Li Y, Wu Y, Huang F, Zhang B (2018) Accelerating convolutional networks via global & dynamic filter pruning.. In: IJCAI, pp 2425\u20132432","DOI":"10.24963\/ijcai.2018\/336"},{"issue":"2","key":"2932_CR30","doi-asserted-by":"publisher","first-page":"574","DOI":"10.1109\/TNNLS.2019.2906563","volume":"31","author":"S Lin","year":"2019","unstructured":"Lin S, Ji R, Li Y, Deng C, Li X (2019) Toward compact convnets via structure-sparsity regularized filter pruning. IEEE transactions on neural networks and learning systems 31(2):574\u2013588","journal-title":"IEEE transactions on neural networks and learning systems"},{"key":"2932_CR31","doi-asserted-by":"crossref","unstructured":"Zhu X, Zhou W, Li H (2018) Improving deep neural network sparsity through decorrelation regularization.. In: IJCAI, pp 3264\u20133270","DOI":"10.24963\/ijcai.2018\/453"},{"key":"2932_CR32","doi-asserted-by":"crossref","unstructured":"Liu C, Wang Y, Han K, Xu C, Xu C (2019) Learning instance-wise sparsity for accelerating deep models. arXiv preprint arXiv:1907.11840","DOI":"10.24963\/ijcai.2019\/416"},{"key":"2932_CR33","doi-asserted-by":"crossref","unstructured":"Huang Z, Wang N (2018) Data-driven sparse structure selection for deep neural networks. In: Proceedings of the European conference on computer vision (ECCV), pp 304\u2013320","DOI":"10.1007\/978-3-030-01270-0_19"},{"key":"2932_CR34","unstructured":"Louizos C, Welling M, Kingma DP (2017) Learning sparse neural networks through l_0 regularization. arXiv preprint arXiv:1712.01312"},{"key":"2932_CR35","unstructured":"Molchanov D, Ashukha A, Vetrov D (2017) Variational dropout sparsifies deep neural networks. In: Proceedings of the 34th International Conference on Machine Learning-Volume 70, JMLR. org, pp 2498\u20132507"},{"key":"2932_CR36","doi-asserted-by":"publisher","first-page":"175703","DOI":"10.1109\/ACCESS.2019.2957203","volume":"7","author":"Z Wang","year":"2019","unstructured":"Wang Z, Lin S, Xie J, Lin Y (2019) Pruning blocks for CNN compression and acceleration via online ensemble distillation. IEEE Access 7:175703\u2013175716","journal-title":"IEEE Access"},{"key":"2932_CR37","doi-asserted-by":"crossref","unstructured":"Lin S, Ji R, Yan C, Zhang B, Cao L, Ye Q, Huang F, Doermann D (2019) Towards optimal structured CNN pruning via generative adversarial learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 2790\u20132799","DOI":"10.1109\/CVPR.2019.00290"},{"key":"2932_CR38","doi-asserted-by":"crossref","unstructured":"Liu N, Ma X, Xu Z, Wang Y, Tang J, Ye J (2020) Autocompress: An automatic DNN structured pruning framework for ultra-high compression rates.. In: AAAI, pp 4876\u20134883","DOI":"10.1609\/aaai.v34i04.5924"},{"key":"2932_CR39","doi-asserted-by":"crossref","unstructured":"He Y, Lin J, Liu Z, Wang H, Li L-J, Han S (2018) Amc: Automl for model compression and acceleration on mobile devices. In: Proceedings of the European Conference on Computer Vision (ECCV), pp 784\u2013800","DOI":"10.1007\/978-3-030-01234-2_48"},{"key":"2932_CR40","unstructured":"Lin C, Zhong Z, Wei W, Yan J (2018) Synaptic strength for convolutional neural network. In: Advances in Neural Information Processing Systems, pp 10149\u201310158"},{"key":"2932_CR41","doi-asserted-by":"crossref","unstructured":"Li Y, Lin S, Zhang B, Liu J, Doermann D, Wu Y, Huang F, Ji R (2019) Exploiting kernel sparsity and entropy for interpretable CNN compression. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 2800\u20132809","DOI":"10.1109\/CVPR.2019.00291"},{"key":"2932_CR42","doi-asserted-by":"crossref","unstructured":"Mao H, Han S, Pool J, Li W, Liu X, Wang Y, Dally WJ (2017) Exploring the regularity of sparse structure in convolutional neural networks. arXiv preprint arXiv:1705.08922","DOI":"10.1109\/CVPRW.2017.241"},{"key":"2932_CR43","unstructured":"Wen W, Wu C, Wang Y, Chen Y, Li H (2016) Learning structured sparsity in deep neural networks. In: Advances in neural information processing systems, pp 2074\u20132082"},{"key":"2932_CR44","unstructured":"Wang H, Zhang Q, Wang Y, Hu R (2018) Structured deep neural network pruning by varying regularization parameters. ArXiv preprint:1804.09461 3"},{"key":"2932_CR45","doi-asserted-by":"crossref","unstructured":"Aubry M, Russell BC (2015) Understanding deep features with computer-generated imagery. In: Proceedings of the IEEE International Conference on Computer Vision, pp 2875\u20132883","DOI":"10.1109\/ICCV.2015.329"},{"key":"2932_CR46","doi-asserted-by":"crossref","unstructured":"Zhang Q, Nian Wu Y, Zhu S-C (2018) Interpretable convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 8827\u20138836","DOI":"10.1109\/CVPR.2018.00920"},{"key":"2932_CR47","doi-asserted-by":"crossref","unstructured":"Wagner J, Kohler JM, Gindele T, Hetzel L, Wiedemer JT, Behnke S (2019) Interpretable and fine-grained visual explanations for convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 9097\u20139107","DOI":"10.1109\/CVPR.2019.00931"},{"key":"2932_CR48","doi-asserted-by":"crossref","unstructured":"Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L-C (2018) Mobilenetv2: Inverted residuals and linear bottlenecks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4510\u20134520","DOI":"10.1109\/CVPR.2018.00474"},{"key":"2932_CR49","unstructured":"Krizhevsky A (2009) Learning multiple layers of features from tiny images. Master\u2019s thesis, University of Tront"},{"key":"2932_CR50","doi-asserted-by":"crossref","unstructured":"Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L (2009) Imagenet: A large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition, Ieee, pp 248\u2013255","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"2932_CR51","unstructured":"Adam P, Sam G, Soumith C, Gregory C, Edward Y, Zachary D, Zeming L, Alban D, Luca A, Adam L (2017) Automatic differentiation in pytorch. In: Proceedings of Neural Information Processing Systems"},{"key":"2932_CR52","unstructured":"Li Z, Gong Y, Ma X, Liu S, Sun M, Zhan Z, Kong Z, Yuan G, Wang Y (2020) SS-Auto: A single-shot, automatic structured weight pruning framework of DNNs with ultra-high efficiency. arXiv preprint arXiv:2001.08839"},{"key":"2932_CR53","doi-asserted-by":"publisher","first-page":"90924","DOI":"10.1109\/ACCESS.2020.2993932","volume":"8","author":"Y Zuo","year":"2020","unstructured":"Zuo Y, Chen B, Shi T, Sun M (2020) Filter pruning without damaging networks capacity. IEEE Access 8:90924\u201390930","journal-title":"IEEE Access"},{"key":"2932_CR54","doi-asserted-by":"publisher","first-page":"293","DOI":"10.1109\/TIP.2020.3035028","volume":"30","author":"G Ding","year":"2020","unstructured":"Ding G, Zhang S, Jia Z, Zhong J, Han J (2020) Where to prune: Using lstm to guide data-dependent soft pruning. IEEE Transactions on Image Processing 30:293\u2013304","journal-title":"IEEE Transactions on Image Processing"},{"key":"2932_CR55","doi-asserted-by":"crossref","unstructured":"Ding X, Ding G, Guo Y, Han J (2019) Centripetal sgd for pruning very deep convolutional networks with complicated structure. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 4943\u20134953","DOI":"10.1109\/CVPR.2019.00508"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-021-02932-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-021-02932-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-021-02932-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T07:27:20Z","timestamp":1655710040000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-021-02932-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,1,14]]},"references-count":55,"journal-issue":{"issue":"9","published-print":{"date-parts":[[2022,7]]}},"alternative-id":["2932"],"URL":"https:\/\/doi.org\/10.1007\/s10489-021-02932-z","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"value":"0924-669X","type":"print"},{"value":"1573-7497","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,1,14]]},"assertion":[{"value":"14 October 2021","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 January 2022","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"<!--Emphasis Type='Bold' removed-->Conflict of Interests"}}]}}