{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,27]],"date-time":"2026-03-27T12:44:57Z","timestamp":1774615497079,"version":"3.50.1"},"reference-count":47,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2023,7,28]],"date-time":"2023-07-28T00:00:00Z","timestamp":1690502400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,7,28]],"date-time":"2023-07-28T00:00:00Z","timestamp":1690502400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities of China","doi-asserted-by":"crossref","award":["JZ2021HGQA0262"],"award-info":[{"award-number":["JZ2021HGQA0262"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61674049 and U19A2053"],"award-info":[{"award-number":["61674049 and U19A2053"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2023,12]]},"DOI":"10.1007\/s00530-023-01139-1","type":"journal-article","created":{"date-parts":[[2023,7,28]],"date-time":"2023-07-28T13:01:47Z","timestamp":1690549307000},"page":"3609-3623","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["A multi-scale feature fusion spatial\u2013channel attention model for background subtraction"],"prefix":"10.1007","volume":"29","author":[{"given":"Yizhong","family":"Yang","sequence":"first","affiliation":[]},{"given":"Tingting","family":"Xia","sequence":"additional","affiliation":[]},{"given":"Dajin","family":"Li","sequence":"additional","affiliation":[]},{"given":"Zhang","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Guangjun","family":"Xie","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,7,28]]},"reference":[{"key":"1139_CR1","unstructured":"Stauffer C., Grimson W.E.L.: Adaptive background mixture models for real-time tracking. In: Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149), Conference Paper pp. 246\u201352 vol. 2, (1999)"},{"key":"1139_CR2","first-page":"751","volume-title":"European Conference on Computer Vision","author":"A Elgammal","year":"2000","unstructured":"Elgammal, A., Harwood, D., Davis, L.: Non-parametric model for background subtraction. In: European Conference on Computer Vision, pp. 751\u2013767. Springer, New York (2000)"},{"key":"1139_CR3","doi-asserted-by":"crossref","unstructured":"Barnich O., Van Droogenbroeck M., Ieee: VIBE: a powerful random technique to estimate the background in video sequences. In: IEEE International Conference on Acoustics, Speech and Signal Processing, Taipei, TAIWAN, 2009, pp. 945\u2013948, (2009).","DOI":"10.1109\/ICASSP.2009.4959741"},{"issue":"4","key":"1139_CR4","doi-asserted-by":"publisher","first-page":"657","DOI":"10.1109\/TPAMI.2006.68","volume":"28","author":"M Heikkila","year":"2006","unstructured":"Heikkila, M., Pietikainen, M.: A texture-based method for modeling the background and detecting moving objects. IEEE Trans. Pattern Anal. Mach. Intell. 28(4), 657\u2013662 (2006)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1139_CR5","doi-asserted-by":"crossref","unstructured":"Braham M., Van Droogenbroeck M.: Deep Background Subtraction with Scene-Specific Convolutional Neural Networks, in 23rd International Conference on Systems, Signals and Image Processing (IWSSIP), Bratislava, SLOVAKIA, 2016, pp. 113\u2013116, (2016)","DOI":"10.1109\/IWSSIP.2016.7502717"},{"key":"1139_CR6","doi-asserted-by":"crossref","unstructured":"Bakkay M. C. et al.: BScGAN: deep background subtraction with conditional generative adversarial networks, in 25th IEEE International Conference on Image Processing (ICIP), Athens, GREECE, 2018, pp. 4018\u20134022, (2018).","DOI":"10.1109\/ICIP.2018.8451603"},{"key":"1139_CR7","doi-asserted-by":"publisher","first-page":"16010","DOI":"10.1109\/ACCESS.2018.2817129","volume":"6","author":"D Zeng","year":"2018","unstructured":"Zeng, D., Zhu, M.: Background subtraction using multiscale fully convolutional network. IEEE Access 6, 16010\u201316021 (2018)","journal-title":"IEEE Access"},{"key":"1139_CR8","doi-asserted-by":"crossref","unstructured":"Braham M., Pierard S., Van Droogenbroeck M.: Semantic background subtraction, in 2017 IEEE International Conference on Image Processing (ICIP), 2017, pp. 4552\u20134556: Ieee.","DOI":"10.1109\/ICIP.2017.8297144"},{"key":"1139_CR9","doi-asserted-by":"publisher","first-page":"635","DOI":"10.1016\/j.patcog.2017.09.040","volume":"76","author":"M Babaee","year":"2018","unstructured":"Babaee, M., Dinh, D.T., Rigoll, G.: A deep convolutional neural network for video sequence background subtraction (in English). Pattern Recogn. 76, 635\u2013649 (2018)","journal-title":"Pattern Recogn."},{"issue":"11","key":"1139_CR10","doi-asserted-by":"publisher","first-page":"2278","DOI":"10.1109\/5.726791","volume":"86","author":"Y Lecun","year":"1998","unstructured":"Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278\u20132324 (1998)","journal-title":"Proc. IEEE"},{"issue":"3","key":"1139_CR11","doi-asserted-by":"publisher","first-page":"1369","DOI":"10.1007\/s10044-019-00845-9","volume":"23","author":"LA Lim","year":"2019","unstructured":"Lim, L.A., Keles, H.Y.: Learning multi-scale features for foreground segmentation. Pattern Anal. Appl. 23(3), 1369\u20131380 (2019)","journal-title":"Pattern Anal. Appl."},{"key":"1139_CR12","unstructured":"Simonyan K., Zisserman A. J. C. S.: Very Deep Convolutional Networks for Large-Scale Image Recognition (2014)"},{"key":"1139_CR13","doi-asserted-by":"crossref","unstructured":"Long et al.: Fully convolutional networks for semantic segmentation, Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2015: 3431\u20133440, (2017).","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"1139_CR14","doi-asserted-by":"crossref","unstructured":"Ronneberger O., Fischer P., Brox T. J. S. I. P.: U-Net: convolutional networks for biomedical image segmentation, International Conference on Medical image computing and computer-assisted intervention. Springer, Cham, 2015, pp. 234\u2013241, (2015).","DOI":"10.1007\/978-3-319-24574-4_28"},{"issue":"10","key":"1139_CR15","doi-asserted-by":"publisher","first-page":"3349","DOI":"10.1109\/TPAMI.2020.2983686","volume":"43","author":"J Wang","year":"2021","unstructured":"Wang, J., et al.: Deep high-resolution representation learning for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 43(10), 3349\u20133364 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"1","key":"1139_CR16","doi-asserted-by":"publisher","first-page":"359","DOI":"10.1109\/TIP.2014.2378053","volume":"24","author":"P-L St-Charles","year":"2015","unstructured":"St-Charles, P.-L., Bilodeau, G.-A., Bergevin, R.: SuBSENSE: a universal change detection method with local adaptive sensitivity. IEEE Trans. Image Process. 24(1), 359\u2013373 (2015)","journal-title":"IEEE Trans. Image Process."},{"issue":"6","key":"1139_CR17","doi-asserted-by":"publisher","first-page":"1527","DOI":"10.1007\/s00371-020-01890-w","volume":"37","author":"RP Singh","year":"2020","unstructured":"Singh, R.P., Sharma, P.: Instance-vote-based motion detection using spatially extended hybrid feature space. Vis. Comput. 37(6), 1527\u20131543 (2020)","journal-title":"Vis. Comput."},{"key":"1139_CR18","doi-asserted-by":"crossref","unstructured":"Zhao X., Wang G., He Z., Liang D., Zhang S., Tan J. J. T. V. C.: Unsupervised inner-point-pairs model for unseen-scene and online moving object detection, pp. 1\u201317, (2022).","DOI":"10.1007\/s00371-022-02417-1"},{"key":"1139_CR19","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.patcog.2022.108719","volume":"129","author":"M Sultana","year":"2022","unstructured":"Sultana, M., Mahmood, A., Jung, S.K.: Unsupervised moving object segmentation using background subtraction and optimal adversarial noise sample search (in English). Pattern Recogn. 129, 11 (2022). (Art. no. 108719)","journal-title":"Pattern Recogn."},{"key":"1139_CR20","doi-asserted-by":"crossref","unstructured":"Cioppa A., Van Droogenbroeck M., Braham M.: Real-time semantic background subtraction, in 2020 IEEE International Conference on Image Processing (ICIP), 2020, pp. 3214\u20133218: IEEE.","DOI":"10.1109\/ICIP40778.2020.9190838"},{"key":"1139_CR21","first-page":"3","volume-title":"Robust Foreground Segmentation in\u00a0RGBD Data from Complex Scenes Using\u00a0Adversarial Networks","author":"M Sultana","year":"2021","unstructured":"Sultana, M., Bouwmans, T., Giraldo, J.H., Jung, S.K.: Robust Foreground Segmentation in\u00a0RGBD Data from Complex Scenes Using\u00a0Adversarial Networks, pp. 3\u201316. Springer International Publishing, Cham (2021)"},{"key":"1139_CR22","doi-asserted-by":"publisher","first-page":"66","DOI":"10.1016\/j.patrec.2016.09.014","volume":"96","author":"Y Wang","year":"2017","unstructured":"Wang, Y., Luo, Z., Jodoin, P.-M.: Interactive deep learning method for segmenting moving objects. Pattern Recogn. Lett. 96, 66\u201375 (2017)","journal-title":"Pattern Recogn. Lett."},{"key":"1139_CR23","doi-asserted-by":"publisher","first-page":"489","DOI":"10.1109\/LSP.2021.3059195","volume":"28","author":"PW Patil","year":"2021","unstructured":"Patil, P.W., Dudhane, A., Murala, S., Gonde, A.B.: Deep adversarial network for scene independent moving object segmentation (in English). IEEE Signal Process. Lett. 28, 489\u2013493 (2021)","journal-title":"IEEE Signal Process. Lett."},{"issue":"3","key":"1139_CR24","doi-asserted-by":"publisher","first-page":"2031","DOI":"10.1109\/TITS.2020.3030801","volume":"23","author":"M Mandal","year":"2022","unstructured":"Mandal, M., Vipparthi, S.K.: Scene independency matters: an empirical study of scene dependent and scene independent evaluation for CNN-based change detection (in English). IEEE Trans. Intell. Transport. Syst. 23(3), 2031\u20132044 (2022)","journal-title":"IEEE Trans. Intell. Transport. Syst."},{"key":"1139_CR25","doi-asserted-by":"publisher","first-page":"546","DOI":"10.1109\/TIP.2020.3037472","volume":"30","author":"M Mandal","year":"2021","unstructured":"Mandal, M., Dhar, V., Mishra, A., Vipparthi, S.K., Abdel-Mottaleb, M.: 3DCD: scene independent end-to-end spatiotemporal feature learning framework for change detection in unseen videos. IEEE Trans. Image Process. 30, 546\u2013558 (2021)","journal-title":"IEEE Trans. Image Process."},{"issue":"12","key":"1139_CR26","doi-asserted-by":"publisher","first-page":"1882","DOI":"10.1109\/LSP.2019.2952253","volume":"26","author":"M Mandal","year":"2019","unstructured":"Mandal, M., Dhar, V., Mishra, A., Vipparthi, S.K.: 3DFR: a swift 3D feature reductionist framework for scene independent change detection. IEEE Signal Process. Lett. 26(12), 1882\u20131886 (2019)","journal-title":"IEEE Signal Process. Lett."},{"key":"1139_CR27","doi-asserted-by":"crossref","unstructured":"Tezcan M. O., Ishwar P., Konrad J., Soc I. C.: BSUV-Net: a fully-convolutional neural network for background subtraction of unseen videos, in IEEE Winter Conference on Applications of Computer Vision (WACV), Snowmass, CO, 2020, pp. 2763\u20132772, 2020.","DOI":"10.1109\/WACV45572.2020.9093464"},{"key":"1139_CR28","doi-asserted-by":"publisher","first-page":"53849","DOI":"10.1109\/ACCESS.2021.3071163","volume":"9","author":"MO Tezcan","year":"2021","unstructured":"Tezcan, M.O., Ishwar, P., Konrad, J.: BSUV-Net 2.0: spatio-temporal data augmentations for video-agnostic supervised background subtraction. IEEE Access 9, 53849\u201353860 (2021)","journal-title":"IEEE Access"},{"key":"1139_CR29","doi-asserted-by":"publisher","first-page":"9058","DOI":"10.1109\/TIP.2021.3122102","volume":"30","author":"J Zhang","year":"2021","unstructured":"Zhang, J., Zhang, X., Zhang, Y., Duan, Y., Li, Y., Pan, Z.: Meta-knowledge learning and domain adaptation for unseen background subtraction. IEEE Trans. Image Process. 30, 9058\u20139068 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"1139_CR30","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2022.103584","volume":"226","author":"I Kajo","year":"2023","unstructured":"Kajo, I., Kas, M., Ruichek, Y., Kamel, N.: Tensor based completion meets adversarial learning: a win-win solution for change detection on unseen videos. Comput. Vis. Image Understand. 226, 103584 (2023)","journal-title":"Comput. Vis. Image Understand."},{"issue":"3","key":"1139_CR31","doi-asserted-by":"publisher","first-page":"395","DOI":"10.1007\/s13735-022-00232-x","volume":"11","author":"I Houhou","year":"2022","unstructured":"Houhou, I., Zitouni, A., Ruichek, Y., Bekhouche, S.E., Kas, M., Taleb-Ahmed, A.: RGBD deep multi-scale network for background subtraction (in English). Int. J. Multimed. Inf. 11(3), 395\u2013407 (2022)","journal-title":"Int. J. Multimed. Inf."},{"key":"1139_CR32","doi-asserted-by":"crossref","unstructured":"Wang Y. et al.: CDnet 2014: an expanded change detection benchmark dataset, in 27th IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, 2014, pp. 393\u2013+, 2014.","DOI":"10.1109\/CVPRW.2014.126"},{"key":"1139_CR33","doi-asserted-by":"publisher","first-page":"103","DOI":"10.1016\/j.cviu.2016.08.005","volume":"152","author":"C Cuevas","year":"2016","unstructured":"Cuevas, C., Y\u00e1\u00f1ez, E.M., Garc\u00eda, N.: Labeled dataset for integral evaluation of moving object detection algorithms: LASIESTA. Comput. Vis. Image Underst. 152, 103\u2013117 (2016)","journal-title":"Comput. Vis. Image Underst."},{"key":"1139_CR34","doi-asserted-by":"crossref","unstructured":"Maddalena L., Petrosino A.: Towards Benchmarking Scene Background Initialization, in 18th International Conference on Image Analysis and Processing (ICIAP), Genoa, ITALY, 2015, vol. 9281, pp. 469\u2013476, 2015.","DOI":"10.1007\/978-3-319-23222-5_57"},{"issue":"5","key":"1139_CR35","doi-asserted-by":"publisher","first-page":"621","DOI":"10.3390\/sym11050621","volume":"11","author":"S-H Lee","year":"2019","unstructured":"Lee, S.-H., Lee, G.-C., Yoo, J., Kwon, S.: WisenetMD: motion detection using dynamic background region analysis. Symmetry 11(5), 621 (2019)","journal-title":"Symmetry"},{"key":"1139_CR36","doi-asserted-by":"crossref","unstructured":"Qi Q. et al.: Background subtraction via regional multi-feature-frequency model in complex scenes (in English). Soft Comput. Article; Early Access p. 14, (2023).","DOI":"10.21203\/rs.3.rs-2275891\/v1"},{"key":"1139_CR37","unstructured":"Chacon-Murguia M. I., Guzman-Pando A.: Moving object detection in video sequences based on a two-frame temporal information CNN (in English), Neural Process. Lett. Article; Early Access p. 25."},{"key":"1139_CR38","doi-asserted-by":"publisher","first-page":"8","DOI":"10.1016\/j.neunet.2019.04.024","volume":"117","author":"T Bouwmans","year":"2019","unstructured":"Bouwmans, T., Javed, S., Sultana, M., Jung, S.K.: Deep neural network concepts for background subtraction: a systematic review and comparative evaluation. Neural Netw. 117, 8\u201366 (2019)","journal-title":"Neural Netw."},{"issue":"9","key":"1139_CR39","doi-asserted-by":"publisher","first-page":"616","DOI":"10.1016\/j.imavis.2013.06.003","volume":"31","author":"C Cuevas","year":"2013","unstructured":"Cuevas, C., Garc\u00eda, N.: Improved background modeling for real-time spatio-temporal non-parametric moving object detection strategies. Image Vis. Comput. 31(9), 616\u2013630 (2013)","journal-title":"Image Vis. Comput."},{"key":"1139_CR40","doi-asserted-by":"crossref","unstructured":"St-Charles P.-L., Bilodeau G.-A., Bergevin R.: A self-adjusting approach to change detection based on background word consensus, in Presented at the 2015 IEEE Winter Conference on Applications of Computer Vision, 2015.","DOI":"10.1109\/WACV.2015.137"},{"key":"1139_CR41","doi-asserted-by":"crossref","unstructured":"Rahmon G., Bunyak F., Seetharaman G., Palaniappan K.: Motion U-Net: multi-cue encoder-decoder network for motion segmentation, in 2020 25th International Conference on Pattern Recognition (ICPR), Conference Paper pp. 8125\u20138132, (2020).","DOI":"10.1109\/ICPR48806.2021.9413211"},{"key":"1139_CR42","doi-asserted-by":"publisher","first-page":"156","DOI":"10.1016\/j.patcog.2017.09.009","volume":"74","author":"D Berj\u00f3n","year":"2018","unstructured":"Berj\u00f3n, D., Cuevas, C., Mor\u00e1n, F., Garc\u00eda, N.: Real-time nonparametric background subtraction with tracking-based foreground update. Pattern Recogn. 74, 156\u2013170 (2018)","journal-title":"Pattern Recogn."},{"issue":"4","key":"1139_CR43","doi-asserted-by":"publisher","first-page":"670","DOI":"10.1109\/TPAMI.2013.239","volume":"36","author":"TSF Haines","year":"2014","unstructured":"Haines, T.S.F., Xiang, T.: Background Subtraction with DirichletProcess Mixture Models. IEEE Trans. Pattern Anal. Mach. Intell. 36(4), 670\u2013683 (2014)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1139_CR44","doi-asserted-by":"crossref","unstructured":"Maddalena L., Petrosino A.: The SOBS algorithm: What are the limits?, in 2012 IEEE computer society conference on computer vision and pattern recognition workshops, 2012, pp. 21\u201326: IEEE.","DOI":"10.1109\/CVPRW.2012.6238922"},{"issue":"7","key":"1139_CR45","doi-asserted-by":"publisher","first-page":"1168","DOI":"10.1109\/TIP.2008.924285","volume":"17","author":"L Maddalena","year":"2008","unstructured":"Maddalena, L., Petrosino, A.: A self-organizing approach to background subtraction for visual surveillance applications. IEEE Trans. Image Process. 17(7), 1168\u20131177 (2008)","journal-title":"IEEE Trans. Image Process."},{"key":"1139_CR46","doi-asserted-by":"publisher","first-page":"2934","DOI":"10.1109\/TIP.2022.3162961","volume":"31","author":"C Zhao","year":"2022","unstructured":"Zhao, C., Hu, K., Basu, A.: Universal background subtraction based on arithmetic distribution neural network. IEEE Trans. Image Process. 31, 2934\u20132949 (2022)","journal-title":"IEEE Trans. Image Process."},{"key":"1139_CR47","doi-asserted-by":"publisher","first-page":"159864","DOI":"10.1109\/ACCESS.2020.3020818","volume":"8","author":"J-Y Kim","year":"2020","unstructured":"Kim, J.-Y., Ha, J.-E.: Foreground objects detection using a fully convolutional network with a background model image and multiple original images. IEEE Access 8, 159864\u2013159878 (2020)","journal-title":"IEEE Access"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-023-01139-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-023-01139-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-023-01139-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,16]],"date-time":"2023-11-16T11:05:21Z","timestamp":1700132721000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-023-01139-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,7,28]]},"references-count":47,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2023,12]]}},"alternative-id":["1139"],"URL":"https:\/\/doi.org\/10.1007\/s00530-023-01139-1","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,7,28]]},"assertion":[{"value":"26 November 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 July 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 July 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}