{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T16:36:55Z","timestamp":1743007015016,"version":"3.40.3"},"publisher-location":"Cham","reference-count":43,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783031113482"},{"type":"electronic","value":"9783031113499"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-11349-9_17","type":"book-chapter","created":{"date-parts":[[2022,7,23]],"date-time":"2022-07-23T17:06:46Z","timestamp":1658596006000},"page":"192-205","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["(MS)2EDNet: Multiscale Motion Saliency Deep Network for Moving Object Detection"],"prefix":"10.1007","author":[{"given":"Santosh Nagnath","family":"Randive","sequence":"first","affiliation":[]},{"given":"Kishor B.","family":"Bhangale","sequence":"additional","affiliation":[]},{"given":"Rahul G.","family":"Mapari","sequence":"additional","affiliation":[]},{"given":"Kiran M.","family":"Napte","sequence":"additional","affiliation":[]},{"given":"Kishor B.","family":"Wane","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,7,24]]},"reference":[{"key":"17_CR1","doi-asserted-by":"crossref","unstructured":"Y. Wang, P. M. Jodoin, F. Porikli, J. Konrad, Y. Benezeth, and P. Ishwar.: An expanded change detection benchmark dataset. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 393\u2013400 (2014)","DOI":"10.1109\/CVPRW.2014.126"},{"key":"17_CR2","doi-asserted-by":"crossref","unstructured":"Toyama, K., Krumm, J., Brumitt, B., Meyers, B.: Wallflower: principles and practice of background maintenance. In: Proceedings of the IEEE International Conference on Computer Vision, vol. 1(1), pp. 255\u2013261 (1999)","DOI":"10.1109\/ICCV.1999.791228"},{"issue":"10","key":"17_CR3","doi-asserted-by":"publisher","first-page":"1174","DOI":"10.3390\/e22101174","volume":"22","author":"AK Gupta","year":"2020","unstructured":"Gupta, A.K., Seal, A., Prasad, M., Khanna, P.: Salient object detection techniques in computer vision\u2014a survey. Entropy 22(10), 1174 (2020)","journal-title":"Entropy"},{"key":"17_CR4","unstructured":"Wu, Z., Dahua Lin, X.T.: Adjustable bounded rectifiers: towards deep binary representations. arXiv Prepr. arXiv:1511.06201, pp. 1\u201311 (2015)"},{"key":"17_CR5","doi-asserted-by":"crossref","unstructured":"Lin, H., Member, S., Liu, T., Chuang, J., Member, S.: Learning a scene background model via classification. IEEE Trans. Signal Process 57(5), 1641\u20131654 (2009)","DOI":"10.1109\/TSP.2009.2014810"},{"key":"17_CR6","doi-asserted-by":"crossref","unstructured":"Agarwala, A., et al.: Interactive digital photomontage. ACM SIGGRAPH 2004 Pap. - SIGGRAPH \u201904 1(212), 294 (2004)","DOI":"10.1145\/1186562.1015718"},{"key":"17_CR7","unstructured":"Xu, X., Huang, T.S.: A loopy belief propagation approach for robust background estimation. In: 26th IEEE Conference on Computer Vision and Pattern Recognition. CVPR (2008)"},{"key":"17_CR8","doi-asserted-by":"crossref","unstructured":"Liang, C.W., Juang, C.F.: Moving object classification using a combination of static appearance features and spatial and temporal entropy values of optical flows. IEEE Trans. Intell. Transpor. Syst. 16(6), 3453\u20133464 (2015)","DOI":"10.1109\/TITS.2015.2459917"},{"key":"17_CR9","unstructured":"Jiang, S., Lu, X.: WeSamBE: a weight-sample-based method for background subtraction. IEEE Trans. Circu. Syst. Video Technol. 8215, 1\u201310 (2017)"},{"key":"17_CR10","doi-asserted-by":"crossref","unstructured":"Xi, T., Zhao, W., Wang, H., Lin, W.: Salient object detection with spatiotemporal background priors for video. IEEE Trans. Image Process. 26(7), 3425\u20133436 (2017)","DOI":"10.1109\/TIP.2016.2631900"},{"key":"17_CR11","unstructured":"Chen, Y., Wang, J., Zhu, B., Tang, M., Lu, H.: Pixel-wise deep sequence learning for moving object detection. IEEE Trans. Circuits Syst. Video Technol. 8215, 1\u201313 (2017)"},{"key":"17_CR12","doi-asserted-by":"crossref","unstructured":"Yeh, C., Member, S., Lin, C., Muchtar, K., Lai, H., Motivation, A.: Three-pronged compensation and hysteresis thresholding for moving object detection in real-time video surveillance. IEEE Trans. Ind. Electron. 64(6), 4945\u20134955 (2017)","DOI":"10.1109\/TIE.2017.2669881"},{"key":"17_CR13","doi-asserted-by":"crossref","unstructured":"Liao, S., Zhao, G., Kellokumpu, V., Pietik\u00e4inen, M., Li, S.Z.: Modeling pixel process with scale invariant local patterns for background subtraction in complex scenes. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 1301\u20131306 (2010)","DOI":"10.1109\/CVPR.2010.5539817"},{"key":"17_CR14","doi-asserted-by":"crossref","unstructured":"Wang, R., Bunyak, F., Seetharaman, G., Palaniappan, K.: Static and moving object detection using flux tensor with split gaussian models. In: IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 420\u2013424 (2014)","DOI":"10.1109\/CVPRW.2014.68"},{"key":"17_CR15","doi-asserted-by":"crossref","unstructured":"St-Charles, P.-L., Bilodeau, G.-A., Bergevin, R.: SuBSENSE: a universal change detection method with local adaptive sensitivity. IEEE Trans. Image Process. 24(1), 359\u2013373 (2015)","DOI":"10.1109\/TIP.2014.2378053"},{"key":"17_CR16","doi-asserted-by":"crossref","unstructured":"Bianco, S., Ciocca, G., Schettini, R.: Combination of video change detection algorithms by genetic programming. IEEE Trans. Evol. Comput. 21(6), 914\u2013928 (2017)","DOI":"10.1109\/TEVC.2017.2694160"},{"key":"17_CR17","doi-asserted-by":"crossref","unstructured":"Romero, J.D., Lado, M.J., Mendez, A.J.: A background modeling and foreground detection algorithm using scaling coefficients defined with a color model called lightness-red-green-blue. IEEE Trans. Image Process. 27(3), 1243\u20131258 (2017)","DOI":"10.1109\/TIP.2017.2776742"},{"key":"17_CR18","doi-asserted-by":"crossref","unstructured":"Lin, Y., Tong, Y., Cao, Y., Zhou, Y., Wang, S.: Visual-attention-based background modeling for detecting infrequently moving objects. IEEE Trans. Circuits Syst. Video Technol. 27(6), 1208\u20131221 (2017)","DOI":"10.1109\/TCSVT.2016.2527258"},{"key":"17_CR19","doi-asserted-by":"crossref","unstructured":"Aytekin, C., Possegger, H., Mauthner, T., Kiranyaz, S., Bischof, H., Gabbouj, M.: Spatiotemporal saliency estimation by spectral foreground detection. IEEE Trans. Multim. 20(1), 82\u201395 (2018)","DOI":"10.1109\/TMM.2017.2713982"},{"key":"17_CR20","unstructured":"Pang, Y., Member, S., Ye, L., Li, X., Pan, J.: Incremental learning with saliency map for moving object detection. IEEE Trans. Circu. Sys. Video Technol. (TCSVT) 1, 1\u201312 (2016)"},{"key":"17_CR21","doi-asserted-by":"crossref","unstructured":"Braham, M., Van Droogenbroeck, M.: Deep background subtraction with scene-specific convolutional neural networks. In: International Conference on Systems, Signals, and Image Processing, pp. 1\u20134 (2016)","DOI":"10.1109\/IWSSIP.2016.7502717"},{"key":"17_CR22","doi-asserted-by":"crossref","unstructured":"Wang, W., Shen, J., Shao, L.: Video salient object detection via fully convolutional networks. IEEE Trans. Image Process. 27(1), 38\u201349 (2018)","DOI":"10.1109\/TIP.2017.2754941"},{"key":"17_CR23","doi-asserted-by":"crossref","unstructured":"Yang, L., Li, J., Member, S., Luo, Y., Member, S.: Deep background modeling using fully convolutional network. IEEE Trans. Intell. Transp. Syst. 19(1), 254\u2013262 (2018)","DOI":"10.1109\/TITS.2017.2754099"},{"key":"17_CR24","unstructured":"Simonyan, K., Zisserman, A.: Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv Prepr. arXiv:1409.1556, pp. 1\u201314 (2014)"},{"key":"17_CR25","doi-asserted-by":"crossref","unstructured":"Badrinarayanan, V., Kendall, A., Cipolla, R.: SegNet: a deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 39(12), 2481\u20132495 (2017)","DOI":"10.1109\/TPAMI.2016.2644615"},{"key":"17_CR26","unstructured":"Roy, S.M., Ghosh, A.: Real-time adaptive histogram min-max bucket (HMMB) model for background subtraction. IEEE Trans. Circuits Syst. Video Technol. 8215(c), 1\u20131 (2017)"},{"key":"17_CR27","doi-asserted-by":"crossref","unstructured":"Cai, B., Xu, X., Jia, K., Qing, C.: DehazeNet\u202f: an end-to-end system for single image haze removal. IEEE Trans. Image Proce.ss 25(11), 1\u201313 (2016)","DOI":"10.1109\/TIP.2016.2598681"},{"key":"17_CR28","unstructured":"Wu, Z., Lin, D., Tang, X.: Adjustable bounded rectifiers: towards deep binary representations. arXiv Prepr. arXiv1511.06201, pp. 1\u201311 (2015)"},{"key":"17_CR29","doi-asserted-by":"crossref","unstructured":"Szegedy, C., et al.: Going deeper with convolutions. In: IEEE Conference on Computer Vision and Pattern Recognition, vol. 07\u201312\u2013June, pp. 1\u20139 (2015)","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"17_CR30","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"17_CR31","doi-asserted-by":"crossref","unstructured":"St-Charles, P.L., Bilodeau, G.A., Bergevin, R.: A self-adjusting approach to change detection based on background word consensus. In: IEEE Winter Conference on Applications of Computer Vision, WACV 2015, pp. 990\u2013997 (2015)","DOI":"10.1109\/WACV.2015.137"},{"key":"17_CR32","doi-asserted-by":"crossref","unstructured":"Babaee, M., Dinh, D.T., Rigoll, G.: A Deep Convolutional Neural Network for Background Subtraction. arXiv preprint arXiv:1702.01731 (2017)","DOI":"10.1016\/j.patcog.2017.09.040"},{"key":"17_CR33","doi-asserted-by":"crossref","unstructured":"Zivkovic, Z., Van Der Heijden, F.: Efficient adaptive density estimation per image pixel for the task of background subtraction. Pattern Recognit. Lett. 27(7), 773\u2013780 (2006)","DOI":"10.1016\/j.patrec.2005.11.005"},{"key":"17_CR34","doi-asserted-by":"crossref","unstructured":"Barnich, O., Van Droogenbroeck, M.: ViBe\u202f: a universal background subtraction algorithm for video sequences. IEEE Trans. Image Process. 20(6), 1709\u20131724 (2011)","DOI":"10.1109\/TIP.2010.2101613"},{"key":"17_CR35","doi-asserted-by":"crossref","unstructured":"Schindler, K., Wang, H.: Smooth foreground-background segmentation for video processing. In: Asian Conference on Computer Visio, pp. 581\u2013590 (2006)","DOI":"10.1007\/11612704_58"},{"key":"17_CR36","doi-asserted-by":"crossref","unstructured":"Pan, J., Li, X., Li, X., Pang, Y.: Incrementally detecting moving objects in video with sparsity and connectivity. Cognit. Comput. 8(3), 420\u2013428 (2016)","DOI":"10.1007\/s12559-015-9373-5"},{"key":"17_CR37","doi-asserted-by":"crossref","unstructured":"He, J., Balzano, L., Szlam, A.: Incremental gradient on the Grassmannian for online foreground and background separation in subsampled video. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1568\u20131575 (2012)","DOI":"10.1109\/CVPR.2012.6247848"},{"key":"17_CR38","doi-asserted-by":"crossref","unstructured":"Guo, X., Wang, X., Yang, L., Cao, X., Ma, Y.: Robust foreground detection using smoothness and arbitrariness constraints. In: European Conference on Computer Vision, pp. 535\u2013550 (2014)","DOI":"10.1007\/978-3-319-10584-0_35"},{"key":"17_CR39","doi-asserted-by":"crossref","unstructured":"Dikmen, M., Huang, T.S.: Robust estimation of foreground in surveillance videos by sparse error estimation. In: 19th International Conference on Pattern Recognition, pp. 1\u20134 (2008)","DOI":"10.1109\/ICPR.2008.4761910"},{"key":"17_CR40","doi-asserted-by":"crossref","unstructured":"Sheikh, Y., Shah, M.: Bayesian modeling of dynamic scenes for object detection. IEEE Trans. Pattern Anal. Mach. Intell. 27(11), 1778\u20131792 (2005)","DOI":"10.1109\/TPAMI.2005.213"},{"key":"17_CR41","doi-asserted-by":"crossref","unstructured":"Kim, K., Chalidabhongse, T.H., Harwood, D., Davis, L.: Real-time foreground-background segmentation using codebook model. Real-Time Imaging 11(3), 172\u2013185 (2005)","DOI":"10.1016\/j.rti.2004.12.004"},{"key":"17_CR42","doi-asserted-by":"crossref","unstructured":"Hofmann, M., Tiefenbacher, P., Rigoll, G.: Background segmentation with feedback: the pixel-based adaptive segmenter. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition Work, pp. 38\u201343 (2012)","DOI":"10.1109\/CVPRW.2012.6238925"},{"key":"17_CR43","doi-asserted-by":"crossref","unstructured":"Yang, L., Cheng, H., Su, J., Li, X.: Pixel-to-model distance for robust background reconstruction. IEEE Trans. Circuits Syst. Video Technol. 26(5), 903\u2013916 (2016)","DOI":"10.1109\/TCSVT.2015.2424052"}],"container-title":["Communications in Computer and Information Science","Computer Vision and Image Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-11349-9_17","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,29]],"date-time":"2024-09-29T19:24:38Z","timestamp":1727637878000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-11349-9_17"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031113482","9783031113499"],"references-count":43,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-11349-9_17","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"24 July 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CVIP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Computer Vision and Image Processing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Rupnagar","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"India","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3 December 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 December 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"cvip2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/iitrpr.cvip2021.com\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"260","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"77","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"20","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"30% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}