{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,13]],"date-time":"2026-01-13T03:33:27Z","timestamp":1768275207986,"version":"3.49.0"},"reference-count":42,"publisher":"Springer Science and Business Media LLC","issue":"8","license":[{"start":{"date-parts":[[2023,6,26]],"date-time":"2023-06-26T00:00:00Z","timestamp":1687737600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,6,26]],"date-time":"2023-06-26T00:00:00Z","timestamp":1687737600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["JZ2021HGQA0262"],"award-info":[{"award-number":["JZ2021HGQA0262"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61674049"],"award-info":[{"award-number":["61674049"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U19A2053"],"award-info":[{"award-number":["U19A2053"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2023,11]]},"DOI":"10.1007\/s11760-023-02645-5","type":"journal-article","created":{"date-parts":[[2023,6,26]],"date-time":"2023-06-26T06:02:21Z","timestamp":1687759341000},"page":"4133-4141","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["A multi-scale inputs and labels model for background subtraction"],"prefix":"10.1007","volume":"17","author":[{"given":"Yizhong","family":"Yang","sequence":"first","affiliation":[]},{"given":"Dajin","family":"Li","sequence":"additional","affiliation":[]},{"given":"Xiang","family":"Li","sequence":"additional","affiliation":[]},{"given":"Zhang","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Guangjun","family":"Xie","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,6,26]]},"reference":[{"key":"2645_CR1","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Fidler, S., Urtasun, R.: Instance-level segmentation for autonomous driving with deep densely connected MRFS. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 669\u2013677 (2016)","DOI":"10.1109\/CVPR.2016.79"},{"key":"2645_CR2","unstructured":"Sen-Ching, S.C., Kamath, C.: Robust techniques for background subtraction in urban traffic video. In: Visual Communications and Image Processing 2004. International Society for Optics and Photonics, vol. 5308, pp. 881-892 (2004)"},{"issue":"4","key":"2645_CR3","doi-asserted-by":"publisher","first-page":"665","DOI":"10.1007\/s11760-013-0576-5","volume":"8","author":"H Wu","year":"2014","unstructured":"Wu, H., Liu, N., Luo, X., et al.: Real-time background subtraction-based video surveillance of people by integrating local texture patterns. SIViP 8(4), 665\u2013676 (2014)","journal-title":"SIViP"},{"issue":"4","key":"2645_CR4","doi-asserted-by":"publisher","first-page":"13-es","DOI":"10.1145\/1177352.1177355","volume":"38","author":"A Yilmaz","year":"2006","unstructured":"Yilmaz, A., Javed, O., Shah, M.: Object tracking: a survey. Acm Comput. Surveys CSUR 38(4), 13-es (2006)","journal-title":"Acm Comput. Surveys CSUR"},{"issue":"1","key":"2645_CR5","first-page":"18","volume":"36","author":"W Li","year":"2013","unstructured":"Li, W., Mahadevan, V., Vasconcelos, N.: Anomaly detection and localization in crowded scenes. IEEE Trans. Pattern Anal. Mach. Intell. 36(1), 18\u201332 (2013)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2645_CR6","unstructured":"Stauffer, C., Grimson, W.E.L.: Adaptive background mixture models for real-time tracking. In: Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149). IEEE, vol. 2 pp. 246\u2013252 (1999)"},{"issue":"6","key":"2645_CR7","doi-asserted-by":"publisher","first-page":"1709","DOI":"10.1109\/TIP.2010.2101613","volume":"20","author":"O Barnich","year":"2011","unstructured":"Barnich, O., Van Droogenbroeck, M.: ViBe: a universal background subtraction algorithm for video sequences. IEEE Trans. Image Proc. 20(6), 1709\u20131724 (2011). https:\/\/doi.org\/10.1109\/TIP.2010.2101613","journal-title":"IEEE Trans. Image Proc."},{"key":"2645_CR8","doi-asserted-by":"crossref","unstructured":"Cho, K., Van Merri\u00ebnboer, B., Gulcehre, C., et al.: Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078, (2014)","DOI":"10.3115\/v1\/D14-1179"},{"key":"2645_CR9","doi-asserted-by":"crossref","unstructured":"Zhang, H., Patel, V.M.: Density-aware single image de-raining using a multi-stream dense network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 695\u2013704 (2018)","DOI":"10.1109\/CVPR.2018.00079"},{"key":"2645_CR10","doi-asserted-by":"publisher","first-page":"354","DOI":"10.1007\/978-3-319-46493-0_22","volume-title":"Computer Vision \u2013 ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, Part IV","author":"Z Cai","year":"2016","unstructured":"Cai, Z., Fan, Q., Feris, R.S., Vasconcelos, N.: A unified multi-scale deep convolutional neural network for fast object detection. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) Computer Vision \u2013 ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, Part IV, pp. 354\u2013370. Springer International Publishing, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46493-0_22"},{"issue":"4","key":"2645_CR11","doi-asserted-by":"publisher","first-page":"617","DOI":"10.1109\/LGRS.2018.2797538","volume":"15","author":"D Zeng","year":"2018","unstructured":"Zeng, D., Zhu, M.: Multiscale fully convolutional network for foreground object detection in infrared videos. IEEE Geosci. Remote Sens. Lett. 15(4), 617\u2013621 (2018)","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"2645_CR12","doi-asserted-by":"publisher","first-page":"66","DOI":"10.1016\/j.patrec.2016.09.014","volume":"96","author":"Y Wang","year":"2017","unstructured":"Wang, Y., Luo, Z., Jodoin, P.M.: Interactive deep learning method for segmenting moving objects. Pattern Recogn. Lett. 96, 66\u201375 (2017)","journal-title":"Pattern Recogn. Lett."},{"key":"2645_CR13","doi-asserted-by":"crossref","unstructured":"Tezcan O, Ishwar P, Konrad J. BSUV-Net: A fully-convolutional neural network for background subtraction of unseen videos[C]\/\/Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision. 2020: 2774\u20132783.","DOI":"10.1109\/WACV45572.2020.9093464"},{"key":"2645_CR14","doi-asserted-by":"publisher","first-page":"546","DOI":"10.1109\/TIP.2020.3037472","volume":"30","author":"M Mandal","year":"2020","unstructured":"Mandal, M., Dhar, V., Mishra, A., et al.: 3DCD: Scene independent end-to-end spatiotemporal feature learning framework for change detection in unseen videos. IEEE Trans. Image Process. 30, 546\u2013558 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"2645_CR15","unstructured":"Sobral, A., Bouwmans, T.: BGS library: A library framework for algorithms evaluation in foreground\/background segmentation. In: Background Modeling and Foreground Detection for Video Surveillance, Boca Raton, FL, USA: CRC Press, ch. 23, pp. 1\u201316 (2014)"},{"key":"2645_CR16","unstructured":"Mandal, M., Vipparthi, S.K.: An empirical review of deep learning frameworks for change detection: model design, experimental frameworks, challenges and research needs. IEEE Trans. Intell. Transp. Syst. 1\u201322 (2021)"},{"key":"2645_CR17","doi-asserted-by":"crossref","unstructured":"Wren, C.: Real-time tracking of the human body. Photonics East, SPIE, 2615 (1995)","DOI":"10.1117\/12.229194"},{"issue":"3","key":"2645_CR18","doi-asserted-by":"publisher","first-page":"467","DOI":"10.1007\/s11760-020-01766-5","volume":"15","author":"H Ahn","year":"2021","unstructured":"Ahn, H., Kang, M.: Dynamic background subtraction with masked RPCA. SIViP 15(3), 467\u2013474 (2021)","journal-title":"SIViP"},{"key":"2645_CR19","doi-asserted-by":"crossref","unstructured":"Braham, M., Van Droogenbroeck, M.: Deep background subtraction with scene-specific convolutional neural networks. In: 2016 International Conference on Systems, Signals and Image Processing (IWSSIP). IEEE, pp. 1\u20134 (2016)","DOI":"10.1109\/IWSSIP.2016.7502717"},{"issue":"2","key":"2645_CR20","doi-asserted-by":"publisher","first-page":"433","DOI":"10.1109\/TCSVT.2018.2795657","volume":"29","author":"TP Nguyen","year":"2018","unstructured":"Nguyen, T.P., Pham, C.C., Ha, S.V.U., et al.: Change detection by training a triplet network for motion feature extraction. IEEE Trans. Circuits Syst. Video Technol. 29(2), 433\u2013446 (2018)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"9","key":"2645_CR21","doi-asserted-by":"publisher","first-page":"2567","DOI":"10.1109\/TCSVT.2017.2770319","volume":"29","author":"Y Chen","year":"2017","unstructured":"Chen, Y., Wang, J., Zhu, B., et al.: Pixelwise deep sequence learning for moving object detection. IEEE Trans. Circuits Syst. Video Technol. 29(9), 2567\u20132579 (2017)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"11","key":"2645_CR22","doi-asserted-by":"publisher","first-page":"4066","DOI":"10.1109\/TITS.2018.2880096","volume":"20","author":"PW Patil","year":"2018","unstructured":"Patil, P.W., Murala, S.: Msfgnet: a novel compact end-to-end deep network for moving object detection. IEEE Trans. Intell. Transp. Syst. 20(11), 4066\u20134077 (2018)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"2645_CR23","doi-asserted-by":"crossref","unstructured":"Minematsu, T., Shimada, A., Taniguchi, R.: Rethinking background and foreground in deep neural network-based background subtraction. In: 2020 IEEE International Conference on Image Processing (ICIP). IEEE, pp. 3229\u20133233 (2020)","DOI":"10.1109\/ICIP40778.2020.9191151"},{"key":"2645_CR24","doi-asserted-by":"crossref","unstructured":"Giraldo, J.H., Bouwmans, T.: Semi-supervised background subtraction of unseen videos: minimization of the total variation of graph signals. In: 2020 IEEE International Conference on Image Processing (ICIP). IEEE, pp. 3224\u20133228 (2020)","DOI":"10.1109\/ICIP40778.2020.9190887"},{"key":"2645_CR25","doi-asserted-by":"publisher","first-page":"31","DOI":"10.1007\/978-3-030-81638-4_3","volume-title":"Frontiers of Computer Vision: 27th International Workshop, IW-FCV 2021, Daegu, South Korea, February 22\u201323, 2021, Revised Selected Papers","author":"JH Giraldo","year":"2021","unstructured":"Giraldo, J.H., Javed, S., Sultana, M., Jung, S.K., Bouwmans, T.: The emerging field of graph signal processing for moving object segmentation. In: Jeong, H., Sumi, K. (eds.) Frontiers of Computer Vision: 27th International Workshop, IW-FCV 2021, Daegu, South Korea, February 22\u201323, 2021, Revised Selected Papers, pp. 31\u201345. Springer International Publishing, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-81638-4_3"},{"key":"2645_CR26","doi-asserted-by":"crossref","unstructured":"Giraldo, J.H., Javed, S., Werghi, N., et al.: Graph CNN for moving object detection in complex environments from unseen videos. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 225\u2013233 (2021)","DOI":"10.1109\/ICCVW54120.2021.00030"},{"key":"2645_CR27","doi-asserted-by":"publisher","first-page":"9058","DOI":"10.1109\/TIP.2021.3122102","volume":"30","author":"J Zhang","year":"2021","unstructured":"Zhang, J., Zhang, X., Zhang, Y., et al.: Meta-knowledge learning and domain adaptation for unseen background subtraction. IEEE Trans. Image Process. 30, 9058\u20139068 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"2645_CR28","doi-asserted-by":"publisher","first-page":"148433","DOI":"10.1109\/ACCESS.2021.3123975","volume":"9","author":"B Hou","year":"2021","unstructured":"Hou, B., Liu, Y., Ling, N., et al.: A fast lightweight 3D separable convolutional neural network with multi-input multi-output for moving object detection. IEEE Access 9, 148433\u2013148448 (2021)","journal-title":"IEEE Access"},{"key":"2645_CR29","doi-asserted-by":"publisher","first-page":"811","DOI":"10.1007\/978-981-16-6328-4_81","volume-title":"Proceedings of 2021 Chinese Intelligent Systems Conference: Volume I","author":"F Huini","year":"2022","unstructured":"Huini, F., Ma, Z., Zhao, B., Yang, Z., Jiang, Y., Zhu, M.: Lightweight convolutional neural network for foreground segmentation. In: Jia, Y., Zhang, W., Yongling, F., Zhiyuan, Y., Zheng, S. (eds.) Proceedings of 2021 Chinese Intelligent Systems Conference: Volume I, pp. 811\u2013819. Springer Singapore, Singapore (2022). https:\/\/doi.org\/10.1007\/978-981-16-6328-4_81"},{"key":"2645_CR30","doi-asserted-by":"crossref","unstructured":"Hou, B., Liu, Y., Ling, N.: A super-fast deep network for moving object detection. In: International Symposium on Circuits and Systems (ISCAS), pp. 1\u20135 (2020)","DOI":"10.1109\/ISCAS45731.2020.9181053"},{"key":"2645_CR31","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 7132\u20137141 (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"2645_CR32","doi-asserted-by":"crossref","unstructured":"Wang, Y., Jodoin, P.M., Porikli, F., et al.: CDnet 2014: An expanded change detection benchmark dataset. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops. pp. 387\u2013394 (2014)","DOI":"10.1109\/CVPRW.2014.126"},{"key":"2645_CR33","doi-asserted-by":"publisher","first-page":"103","DOI":"10.1016\/j.cviu.2016.08.005","volume":"152","author":"C Cuevas","year":"2016","unstructured":"Cuevas, C., Y\u00e1\u00f1ez, E.M., Garc\u00eda, N.: Labeled dataset for integral evaluation of moving object detection algorithms: LASIESTA. Comput. Vis. Image Underst. 152, 103\u2013117 (2016)","journal-title":"Comput. Vis. Image Underst."},{"key":"2645_CR34","doi-asserted-by":"publisher","first-page":"359","DOI":"10.1109\/TIP.2014.2378053","volume":"24","author":"PL St-Charles","year":"2015","unstructured":"St-Charles, P.L., Bilodeau, G.A., Bergevin, R.: SuBSENSE: a universal change detection method with local adaptive sensitivity. IEEE Trans. Image Proc. 24, 359\u2013373 (2015)","journal-title":"IEEE Trans. Image Proc."},{"issue":"6","key":"2645_CR35","doi-asserted-by":"publisher","first-page":"914","DOI":"10.1109\/TEVC.2017.2694160","volume":"21","author":"S Bianco","year":"2017","unstructured":"Bianco, S., Ciocca, G., Schettini, R.: Combination of video change detection algorithms by genetic programming. IEEE Trans. Evol. Comput. 21(6), 914\u2013928 (2017)","journal-title":"IEEE Trans. Evol. Comput."},{"key":"2645_CR36","doi-asserted-by":"publisher","first-page":"43450","DOI":"10.1109\/ACCESS.2018.2861223","volume":"6","author":"Z Hu","year":"2018","unstructured":"Hu, Z., Turki, T., Phan, N., et al.: A 3D atrous convolutional long short-term memory network for background subtraction. IEEE Access 6, 43450\u201343459 (2018)","journal-title":"IEEE Access"},{"issue":"7","key":"2645_CR37","doi-asserted-by":"publisher","first-page":"4818","DOI":"10.1109\/TII.2020.3017078","volume":"17","author":"A Shahbaz","year":"2020","unstructured":"Shahbaz, A., Jo, K.H.: Deep atrous spatial features-based supervised foreground detection algorithm for industrial surveillance systems. IEEE Trans. Industr. Inf. 17(7), 4818\u20134826 (2020)","journal-title":"IEEE Trans. Industr. Inf."},{"key":"2645_CR38","doi-asserted-by":"publisher","first-page":"127515","DOI":"10.1109\/ACCESS.2021.3111686","volume":"9","author":"JY Kim","year":"2021","unstructured":"Kim, J.Y., Ha, J.E.: Generation of background model image using foreground model. IEEE Access 9, 127515\u2013127530 (2021)","journal-title":"IEEE Access"},{"issue":"9","key":"2645_CR39","doi-asserted-by":"publisher","first-page":"616","DOI":"10.1016\/j.imavis.2013.06.003","volume":"31","author":"C Cuevas","year":"2013","unstructured":"Cuevas, C., Garc\u00eda, N.: Improved background modeling for real-time spatio-temporal non-parametric moving object detection strategies. Image Vis. Comput. 31(9), 616\u2013630 (2013)","journal-title":"Image Vis. Comput."},{"key":"2645_CR40","doi-asserted-by":"crossref","unstructured":"Maddalena, L., Petrosino, A.: The SOBS algorithm: What are the limits? In: 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops. IEEE, pp. 21\u201326 (2012)","DOI":"10.1109\/CVPRW.2012.6238922"},{"issue":"4","key":"2645_CR41","doi-asserted-by":"publisher","first-page":"670","DOI":"10.1109\/TPAMI.2013.239","volume":"36","author":"TSF Haines","year":"2013","unstructured":"Haines, T.S.F., Xiang, T.: Background subtraction with dirichletprocess mixture models. IEEE Trans. Pattern Anal. Mach. Intell. 36(4), 670\u2013683 (2013)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2645_CR42","doi-asserted-by":"publisher","first-page":"156","DOI":"10.1016\/j.patcog.2017.09.009","volume":"74","author":"D Berj\u00f3n","year":"2018","unstructured":"Berj\u00f3n, D., Cuevas, C., Mor\u00e1n, F., et al.: Real-time nonparametric background subtraction with tracking-based foreground update. Pattern Recogn. 74, 156\u2013170 (2018)","journal-title":"Pattern Recogn."}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-023-02645-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-023-02645-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-023-02645-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,23]],"date-time":"2024-10-23T01:39:12Z","timestamp":1729647552000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-023-02645-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,26]]},"references-count":42,"journal-issue":{"issue":"8","published-print":{"date-parts":[[2023,11]]}},"alternative-id":["2645"],"URL":"https:\/\/doi.org\/10.1007\/s11760-023-02645-5","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"value":"1863-1703","type":"print"},{"value":"1863-1711","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,6,26]]},"assertion":[{"value":"14 May 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 December 2022","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 May 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 June 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no Competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"Not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"Not applicable.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}}]}}