{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,4]],"date-time":"2026-04-04T17:55:29Z","timestamp":1775325329848,"version":"3.50.1"},"publisher-location":"Cham","reference-count":53,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031200823","type":"print"},{"value":"9783031200830","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-20083-0_38","type":"book-chapter","created":{"date-parts":[[2022,11,2]],"date-time":"2022-11-02T19:46:34Z","timestamp":1667418394000},"page":"641-657","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["Soft Masking for\u00a0Cost-Constrained Channel Pruning"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-5854-9117","authenticated-orcid":false,"given":"Ryan","family":"Humble","sequence":"first","affiliation":[]},{"given":"Maying","family":"Shen","sequence":"additional","affiliation":[]},{"given":"Jorge Albericio","family":"Latorre","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1938-3836","authenticated-orcid":false,"given":"Eric","family":"Darve","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7535-6322","authenticated-orcid":false,"given":"Jose","family":"Alvarez","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,3]]},"reference":[{"key":"38_CR1","unstructured":"Alvarez, J.M., Salzmann, M.: Learning the number of neurons in deep networks. In: NeurIPS, pp. 2262\u20132270 (2016)"},{"key":"38_CR2","unstructured":"Bengio, Y., L\u00e9onard, N., Courville, A.C.: Estimating or propagating gradients through stochastic neurons for conditional computation. CoRR abs\/1308.3432 (2013)"},{"key":"38_CR3","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"409","DOI":"10.1007\/978-3-030-01237-3_25","volume-title":"Computer Vision \u2013 ECCV 2018","author":"C Chen","year":"2018","unstructured":"Chen, C., Tung, F., Vedula, N., Mori, G.: Constraint-aware deep neural network compression. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11212, pp. 409\u2013424. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01237-3_25"},{"key":"38_CR4","unstructured":"Chetlur, S., et al.: CUDNN: efficient primitives for deep learning. CoRR abs\/1410.0759 (2014)"},{"key":"38_CR5","doi-asserted-by":"crossref","unstructured":"Dai, X., et al.: ChamNet: towards efficient network design through platform-aware model adaptation. In: CVPR, pp. 11398\u201311407 (2019)","DOI":"10.1109\/CVPR.2019.01166"},{"key":"38_CR6","unstructured":"Dettmers, T., Zettlemoyer, L.: Sparse networks from scratch: faster training without losing performance. CoRR abs\/1907.04840 (2019)"},{"key":"38_CR7","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"540","DOI":"10.1007\/978-3-030-01252-6_32","volume-title":"Computer Vision \u2013 ECCV 2018","author":"J-D Dong","year":"2018","unstructured":"Dong, J.-D., Cheng, A.-C., Juan, D.-C., Wei, W., Sun, M.: DPP-Net: device-aware progressive search for pareto-optimal neural architectures. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11215, pp. 540\u2013555. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01252-6_32"},{"key":"38_CR8","unstructured":"Evci, U., Gale, T., Menick, J., Castro, P.S., Elsen, E.: Rigging the lottery: making all tickets winners. In: ICML, pp. 2943\u20132952 (2020)"},{"issue":"2","key":"38_CR9","doi-asserted-by":"publisher","first-page":"303","DOI":"10.1007\/s11263-009-0275-4","volume":"88","author":"M Everingham","year":"2010","unstructured":"Everingham, M., Gool, L.V., Williams, C.K.I., Winn, J.M., Zisserman, A.: The pascal visual object classes (VOC) challenge. Int. J. Comput. Vis. 88(2), 303\u2013338 (2010)","journal-title":"Int. J. Comput. Vis."},{"key":"38_CR10","unstructured":"Guo, Y., Yao, A., Chen, Y.: Dynamic network surgery for efficient DNNs. In: Lee, D.D., Sugiyama, M., von Luxburg, U., Guyon, I., Garnett, R. (eds.) NeurIPS, pp. 1379\u20131387 (2016)"},{"key":"38_CR11","unstructured":"Han, S., Mao, H., Dally, W.J.: Deep compression: compressing deep neural network with pruning, trained quantization and huffman coding. In: ICLR (2016)"},{"key":"38_CR12","unstructured":"Han, S., Pool, J., Tran, J., Dally, W.J.: Learning both weights and connections for efficient neural network. In: NeurIPS, pp. 1135\u20131143 (2015)"},{"key":"38_CR13","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"38_CR14","doi-asserted-by":"crossref","unstructured":"He, Y., Kang, G., Dong, X., Fu, Y., Yang, Y.: Soft filter pruning for accelerating deep convolutional neural networks. In: IJCAI, pp. 2234\u20132240 (2018)","DOI":"10.24963\/ijcai.2018\/309"},{"key":"38_CR15","doi-asserted-by":"crossref","unstructured":"He, Y., Zhang, X., Sun, J.: Channel pruning for accelerating very deep neural networks. In: ICCV, pp. 1398\u20131406 (2017)","DOI":"10.1109\/ICCV.2017.155"},{"key":"38_CR16","unstructured":"Howard, A.G., et al.: MobileNets: efficient convolutional neural networks for mobile vision applications. CoRR abs\/1704.04861 (2017)"},{"key":"38_CR17","unstructured":"de Jorge, P., Sanyal, A., Behl, H.S., Torr, P.H.S., Rogez, G., Dokania, P.K.: Progressive skeletonization: trimming more fat from a network at initialization. In: ICLR (2021)"},{"key":"38_CR18","unstructured":"Kang, M., Han, B.: Operation-aware soft channel pruning using differentiable masks. In: ICML, pp. 5122\u20135131 (2020)"},{"key":"38_CR19","unstructured":"Kusupati, A., et al.: Soft threshold weight reparameterization for learnable sparsity. In: ICML, pp. 5544\u20135555 (2020)"},{"key":"38_CR20","unstructured":"LeCun, Y., Denker, J.S., Solla, S.A.: Optimal brain damage. In: NeurIPS, pp. 598\u2013605 (1989)"},{"key":"38_CR21","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"639","DOI":"10.1007\/978-3-030-58536-5_38","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Li","year":"2020","unstructured":"Li, B., Wu, B., Su, J., Wang, G.: EagleEye: fast sub-net evaluation for\u00a0efficient neural network pruning. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12347, pp. 639\u2013654. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58536-5_38"},{"key":"38_CR22","unstructured":"Li, H., Kadav, A., Durdanovic, I., Samet, H., Graf, H.P.: Pruning filters for efficient convnets. In: ICLR (2017)"},{"key":"38_CR23","unstructured":"Lin, T., Stich, S.U., Barba, L., Dmitriev, D., Jaggi, M.: Dynamic model pruning with feedback. In: ICLR (2020)"},{"key":"38_CR24","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"21","DOI":"10.1007\/978-3-319-46448-0_2","volume-title":"Computer Vision \u2013 ECCV 2016","author":"W Lui","year":"2016","unstructured":"Lui, W., et al.: SSD: single shot MultiBox detector. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 21\u201337. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46448-0_2"},{"key":"38_CR25","doi-asserted-by":"crossref","unstructured":"Liu, Z., Li, J., Shen, Z., Huang, G., Yan, S., Zhang, C.: Learning efficient convolutional networks through network slimming. In: ICCV, pp. 2755\u20132763 (2017)","DOI":"10.1109\/ICCV.2017.298"},{"key":"38_CR26","unstructured":"Liu, Z., Sun, M., Zhou, T., Huang, G., Darrell, T.: Rethinking the value of network pruning. In: ICLR (2019)"},{"key":"38_CR27","doi-asserted-by":"crossref","unstructured":"Luo, J., Wu, J., Lin, W.: ThiNet: a filter level pruning method for deep neural network compression. In: ICCV, pp. 5068\u20135076 (2017)","DOI":"10.1109\/ICCV.2017.541"},{"key":"38_CR28","unstructured":"Mishra, A.K., et al.: Accelerating sparse deep neural networks. CoRR abs\/2104.08378 (2021)"},{"key":"38_CR29","doi-asserted-by":"crossref","unstructured":"Molchanov, P., Mallya, A., Tyree, S., Frosio, I., Kautz, J.: Importance estimation for neural network pruning. In: CVPR, pp. 11264\u201311272 (2019)","DOI":"10.1109\/CVPR.2019.01152"},{"key":"38_CR30","unstructured":"Molchanov, P., Tyree, S., Karras, T., Aila, T., Kautz, J.: Pruning convolutional neural networks for resource efficient inference. In: ICLR (2017)"},{"key":"38_CR31","unstructured":"Mostafa, H., Wang, X.: Parameter efficient training of deep convolutional neural networks by dynamic sparse reparameterization. In: ICML, pp. 4646\u20134655 (2019)"},{"key":"38_CR32","unstructured":"NVIDIA deep learning performance guide: convolutional layers user guide. https:\/\/docs.nvidia.com\/deeplearning\/performance\/dl-performance-convolutional\/index.html. Accessed 15 Nov 2021"},{"key":"38_CR33","unstructured":"NVIDIA Deep Learning Examples: ResNet50 v1.5 For Pytorch. https:\/\/github.com\/NVIDIA\/DeepLearningExamples\/blob\/master\/PyTorch\/Classification\/ConvNets\/resnet50v1.5\/README.md. Accessed 15 Nov 2021"},{"key":"38_CR34","unstructured":"Paszke, A., et al.: PyTorch: an imperative style, high-performance deep learning library. In: 2019 Wallach, H.M., Larochelle, H., Beygelzimer, A., d\u2019Alch\u00e9-Buc, F., Fox, E.B., Garnett, R. (eds.) NeurIPS, pp. 8024\u20138035 (2019)"},{"key":"38_CR35","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"525","DOI":"10.1007\/978-3-319-46493-0_32","volume-title":"Computer Vision \u2013 ECCV 2016","author":"M Rastegari","year":"2016","unstructured":"Rastegari, M., Ordonez, V., Redmon, J., Farhadi, A.: XNOR-Net: imagenet classification using binary convolutional neural networks. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9908, pp. 525\u2013542. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46493-0_32"},{"issue":"3","key":"38_CR36","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O., et al.: Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 115(3), 211\u2013252 (2015)","journal-title":"Int. J. Comput. Vis."},{"key":"38_CR37","unstructured":"Shen, M., Yin, H., Molchanov, P., Mao, L., Liu, J., Alvarez, J.M.: HALP: hardware-aware latency pruning. CoRR abs\/2110.10811 (2021)"},{"issue":"3","key":"38_CR38","doi-asserted-by":"publisher","first-page":"503","DOI":"10.1287\/opre.27.3.503","volume":"27","author":"P Sinha","year":"1979","unstructured":"Sinha, P., Zoltners, A.A.: The multiple-choice knapsack problem. Oper. Res. 27(3), 503\u2013515 (1979)","journal-title":"Oper. Res."},{"key":"38_CR39","unstructured":"Stosic, D., Stosic, D.: Search spaces for neural model training. CoRR abs\/2105.12920 (2021)"},{"key":"38_CR40","doi-asserted-by":"crossref","unstructured":"Su, X., You, S., Wang, F., Qian, C., Zhang, C., Xu, C.: BCNet: searching for network width with bilaterally coupled network. In: CVPR, pp. 2175\u20132184 (2021)","DOI":"10.1109\/CVPR46437.2021.00221"},{"key":"38_CR41","doi-asserted-by":"crossref","unstructured":"Tan, M., et al.: MnasNet: platform-aware neural architecture search for mobile. In: CVPR, pp. 2820\u20132828 (2019)","DOI":"10.1109\/CVPR.2019.00293"},{"key":"38_CR42","unstructured":"Wang, H., Qin, C., Zhang, Y., Fu, Y.: Neural pruning via growing regularization. In: ICLR (2021)"},{"key":"38_CR43","unstructured":"Wortsman, M., Farhadi, A., Rastegari, M.: Discovering neural wirings. In: NeurIPS, pp. 2680\u20132690 (2019)"},{"key":"38_CR44","doi-asserted-by":"crossref","unstructured":"Wu, Y., Liu, C., Chen, B., Chien, S.: Constraint-aware importance estimation for global filter pruning under multiple resource constraints. In: CVPR, pp. 2935\u20132943 (2020)","DOI":"10.1109\/CVPRW50498.2020.00351"},{"key":"38_CR45","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"289","DOI":"10.1007\/978-3-030-01249-6_18","volume-title":"Computer Vision \u2013 ECCV 2018","author":"T-J Yang","year":"2018","unstructured":"Yang, T.-J., et al.: NetAdapt: platform-aware neural network adaptation for mobile applications. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11214, pp. 289\u2013304. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01249-6_18"},{"key":"38_CR46","doi-asserted-by":"crossref","unstructured":"Yang, T., Liao, Y., Sze, V.: Netadaptv2: efficient neural architecture search with fast super-network training and architecture optimization. In: CVPR, pp. 2402\u20132411. Computer Vision Foundation\/IEEE (2021)","DOI":"10.1109\/CVPR46437.2021.00243"},{"key":"38_CR47","unstructured":"Ye, J., Lu, X., Lin, Z., Wang, J.Z.: Rethinking the smaller-norm-less-informative assumption in channel pruning of convolution layers. In: ICLR (2018)"},{"key":"38_CR48","doi-asserted-by":"crossref","unstructured":"Yin, H., et al.: Dreaming to distill: data-free knowledge transfer via DeepInversion. In: CVPR, pp. 8712\u20138721 (2020)","DOI":"10.1109\/CVPR42600.2020.00874"},{"key":"38_CR49","unstructured":"You, Z., Yan, K., Ye, J., Ma, M., Wang, P.: Gate decorator: global filter pruning method for accelerating deep convolutional neural networks. In: NeurIPS, pp. 2130\u20132141 (2019)"},{"key":"38_CR50","unstructured":"Yu, J., Huang, T.S.: Network slimming by slimmable networks: towards one-shot architecture search for channel numbers. CoRR abs\/1903.11728 (2019)"},{"key":"38_CR51","doi-asserted-by":"crossref","unstructured":"Yu, R., et al.: NISP: pruning networks using neuron importance score propagation. In: CVPR, pp. 9194\u20139203 (2018)","DOI":"10.1109\/CVPR.2018.00958"},{"key":"38_CR52","unstructured":"Zhang, C., Bengio, S., Hardt, M., Recht, B., Vinyals, O.: Understanding deep learning requires rethinking generalization. In: ICLR (2017)"},{"key":"38_CR53","unstructured":"Zhou, A., et al.: Learning N: M fine-grained structured sparse neural networks from scratch. In: ICLR (2021)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-20083-0_38","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T12:11:19Z","timestamp":1710331879000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-20083-0_38"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031200823","9783031200830"],"references-count":53,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-20083-0_38","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"3 November 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}