{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T22:24:38Z","timestamp":1742941478573,"version":"3.40.3"},"publisher-location":"Cham","reference-count":30,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031490170"},{"type":"electronic","value":"9783031490187"}],"license":[{"start":{"date-parts":[[2023,11,27]],"date-time":"2023-11-27T00:00:00Z","timestamp":1701043200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,11,27]],"date-time":"2023-11-27T00:00:00Z","timestamp":1701043200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-49018-7_5","type":"book-chapter","created":{"date-parts":[[2023,11,26]],"date-time":"2023-11-26T23:02:21Z","timestamp":1701039741000},"page":"63-75","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["History Based Incremental Singular Value Decomposition for Background Initialization and Foreground Segmentation"],"prefix":"10.1007","author":[{"given":"Ibrahim","family":"Kajo","sequence":"first","affiliation":[]},{"given":"Yassine","family":"Ruichek","sequence":"additional","affiliation":[]},{"given":"Nidal","family":"Kamel","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,11,27]]},"reference":[{"key":"5_CR1","doi-asserted-by":"crossref","unstructured":"Bouwmans, T., Porikli, F., H\u00f6ferlin, B., Vacavant, A.: Background Modeling and Foreground Detection for Video Surveillance. Chapman and Hall\/CRC, Boca Raton (2014)","DOI":"10.1201\/b17223"},{"key":"5_CR2","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"469","DOI":"10.1007\/978-3-319-23222-5_57","volume-title":"New Trends in Image Analysis and Processing -- ICIAP 2015 Workshops","author":"L Maddalena","year":"2015","unstructured":"Maddalena, L., Petrosino, A.: Towards benchmarking scene background initialization. In: Murino, V., Puppo, E., Sona, D., Cristani, M., Sansone, C. (eds.) ICIAP 2015. LNCS, vol. 9281, pp. 469\u2013476. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-23222-5_57"},{"key":"5_CR3","doi-asserted-by":"crossref","unstructured":"Bouwmans, T., Maddalena, L., Petrosino, A.: Scene background initialization: a taxonomy. Pattern Recognit Lett. 96, 3\u201311 (2017)","DOI":"10.1016\/j.patrec.2016.12.024"},{"key":"5_CR4","doi-asserted-by":"crossref","unstructured":"Javed, S., Mahmood, A., Bouwmans, T., Jung, S.K.: Spatiotemporal low-rank modeling for complex scene background initialization. IEEE Trans. Circuits Syst. Video Technol. 28, 1315\u20131329 (2018)","DOI":"10.1109\/TCSVT.2016.2632302"},{"key":"5_CR5","doi-asserted-by":"crossref","unstructured":"Kajo, I., Kamel, N., Ruichek, Y.: Self-motion-assisted tensor completion method for background initialization in complex video sequences. IEEE Trans. Image Process. 29, 1915\u20131928 (2019)","DOI":"10.1109\/TIP.2019.2946098"},{"key":"5_CR6","doi-asserted-by":"crossref","unstructured":"Cuevas, C., Mart\u00ednez, R., Garc\u00eda, N.: Detection of stationary foreground objects: a survey. Comput. Vis. Image Underst. 152, 41\u201357 (2016)","DOI":"10.1016\/j.cviu.2016.07.001"},{"key":"5_CR7","doi-asserted-by":"crossref","unstructured":"Kajo, I., Kamel, N., Ruichek, Y.: Incremental tensor-based completion method for detection of stationary foreground objects. IEEE Trans. Circuits Syst. Video Technol. 29, 1325\u20131338 (2019)","DOI":"10.1109\/TCSVT.2018.2841825"},{"key":"5_CR8","doi-asserted-by":"crossref","unstructured":"Ingersoll, K., Niedfeldt, P.C., Beard, R.W.: Multiple target tracking and stationary object detection in video with Recursive-RANSAC and tracker-sensor feedback. In: 2015 International Conference on Unmanned Aircraft Systems (ICUAS), pp. 1320\u20131329. IEEE (2015)","DOI":"10.1109\/ICUAS.2015.7152426"},{"key":"5_CR9","doi-asserted-by":"crossref","unstructured":"Ortego, D., SanMiguel, J.C.: Multi-feature stationary foreground detection for crowded video-surveillance. In: 2014 IEEE International Conference on Image Processing (ICIP), pp. 2403\u20132407. IEEE (2014)","DOI":"10.1109\/ICIP.2014.7025486"},{"key":"5_CR10","doi-asserted-by":"publisher","first-page":"1208","DOI":"10.1109\/TCSVT.2016.2527258","volume":"27","author":"Y Lin","year":"2017","unstructured":"Lin, Y., Tong, Y., Cao, Y., Zhou, Y., Wang, S.: Visual-attention-based background modeling for detecting infrequently moving objects. IEEE Trans. Circuits Syst. Video Technol. 27, 1208\u20131221 (2017)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"5_CR11","doi-asserted-by":"crossref","unstructured":"Kim, J., Kang, B.: Nonparametric state machine with multiple features for abnormal object classification. In: IEEE Proceedings of 2014 11th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), pp. 199\u2013203. Springer (2014)","DOI":"10.1109\/AVSS.2014.6918668"},{"key":"5_CR12","doi-asserted-by":"crossref","unstructured":"Tian, Y., Feris, R.S., Liu, H., Hampapur, A., Sun, M.-T.: Robust detection of abandoned and removed objects in complex surveillance videos. IEEE Trans. Syst. Man Cybern. C. 41, 565\u2013576 (2011)","DOI":"10.1109\/TSMCC.2010.2065803"},{"key":"5_CR13","doi-asserted-by":"crossref","unstructured":"Thomaz, L.A., da Silva, A.F., da Silva, E.A.B., Netto, S.L., Bian, X., Krim, H.: Abandoned object detection using operator-space pursuit. In: IEEE Proceedings of IEEE International Conference on Image Processing (ICIP), Springer, pp. 1980\u20131984 (2015)","DOI":"10.1109\/ICIP.2015.7351147"},{"key":"5_CR14","doi-asserted-by":"crossref","unstructured":"Muchtar, K., Lin, C.-Y., Kang, L.-W., Yeh, C.-H.: Abandoned object detection in complicated environments. In: IEEE Proceedings of 2013 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, pp. 1\u20136. Springer (2013)","DOI":"10.1109\/APSIPA.2013.6694206"},{"key":"5_CR15","doi-asserted-by":"crossref","unstructured":"Pan, J., Fan, Q., Pankanti, S.: Robust abandoned object detection using region-level analysis. In: IEEE Proceedings of 18th IEEE International Conference on Image Processing, pp. 3597\u20133600. Springer (2011)","DOI":"10.1109\/ICIP.2011.6116495"},{"key":"5_CR16","doi-asserted-by":"crossref","unstructured":"Porikli, F., Ivanov, Y., Haga, T.: Robust abandoned object detection using dual foregrounds. EURASIP J. Adv. Signal Process. 2008 (2008)","DOI":"10.1155\/2008\/197875"},{"key":"5_CR17","doi-asserted-by":"crossref","unstructured":"Tezcan, M.O., Ishwar, P., Konrad, J.: BSUV-Net: a fully-convolutional neural network for background subtraction of unseen videos. In: IEEE, Proceedings of 2020 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 2763\u20132772. Springer (2020)","DOI":"10.1109\/WACV45572.2020.9093464"},{"key":"5_CR18","doi-asserted-by":"crossref","unstructured":"Tezcan, M.O., Ishwar, P., Konrad, J.: BSUV-Net 2.0: spatio-temporal data augmentations for video-agnostic supervised background subtraction. arXiv preprint arXiv:2101.09585 (2021)","DOI":"10.1109\/ACCESS.2021.3071163"},{"key":"5_CR19","unstructured":"Kajo, I., Kamel, N., Ruichek, Y.: Tensor-based approach for background-foreground separation in maritime sequences. IEEE Trans. Intell. Transport. Syst., 1\u201314 (2020)"},{"key":"5_CR20","doi-asserted-by":"publisher","first-page":"178","DOI":"10.1016\/j.neucom.2019.04.088","volume":"394","author":"W Zheng","year":"2020","unstructured":"Zheng, W., Wang, K., Wang, F.-Y.: A novel background subtraction algorithm based on parallel vision and Bayesian GANs. Neurocomputing 394, 178\u2013200 (2020)","journal-title":"Neurocomputing"},{"key":"5_CR21","doi-asserted-by":"publisher","first-page":"4066","DOI":"10.1109\/TITS.2018.2880096","volume":"20","author":"PW Patil","year":"2019","unstructured":"Patil, P.W., Murala, S.: MSFgNet: a novel compact end-to-end deep network for moving object detection. IEEE Trans. Intell. Transport. Syst. 20, 4066\u20134077 (2019)","journal-title":"IEEE Trans. Intell. Transport. Syst."},{"key":"5_CR22","unstructured":"Mandal, M., Vipparthi, S.K.: Scene independency matters: an empirical study of scene dependent and scene independent evaluation for CNN-based change detection. IEEE Trans. Intell. Transport. Syst., 1\u201314 (2020)"},{"key":"5_CR23","doi-asserted-by":"crossref","unstructured":"Mandal, M., Dhar, V., Mishra, A., Vipparthi, S.K., Abdel-Mottaleb, M.: 3DCD: scene independent end-to-end spatiotemporal feature learning framework for change detection in unseen videos. IEEE Trans. Image Process. 30, 546\u2013558 (2021)","DOI":"10.1109\/TIP.2020.3037472"},{"key":"5_CR24","doi-asserted-by":"crossref","unstructured":"Levey, A., Lindenbaum, M.: Sequential Karhunen-Loeve basis extraction and its application to images. IEEE Trans. Image Process. 9, 1371\u20131374 (2000)","DOI":"10.1109\/83.855432"},{"key":"5_CR25","doi-asserted-by":"publisher","first-page":"125","DOI":"10.1007\/s11263-007-0075-7","volume":"77","author":"DA Ross","year":"2008","unstructured":"Ross, D.A., Lim, J., Lin, R.-S., Yang, M.-H.: Incremental learning for robust visual tracking. Int. J. Comput. Vis. 77, 125\u2013141 (2008)","journal-title":"Int. J. Comput. Vis."},{"key":"5_CR26","doi-asserted-by":"crossref","unstructured":"Goyette, N., Jodoin, P.-M., Porikli, F., Konrad, J., Ishwar, P.: Changedetection.net: a new change detection benchmark dataset. In: 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, pp. 1\u20138. IEEE (2012)","DOI":"10.1109\/CVPRW.2012.6238919"},{"key":"5_CR27","doi-asserted-by":"publisher","first-page":"103","DOI":"10.1016\/j.cviu.2016.08.005","volume":"152","author":"C Cuevas","year":"2016","unstructured":"Cuevas, C., Y\u00e1\u00f1ez, E.M., Garc\u00eda, N.: Labeled dataset for integral evaluation of moving object detection algorithms: LASIESTA. Comput. Vis. Image Underst. 152, 103\u2013117 (2016)","journal-title":"Comput. Vis. Image Underst."},{"key":"5_CR28","doi-asserted-by":"publisher","first-page":"256","DOI":"10.1016\/j.patrec.2018.08.002","volume":"112","author":"LA Lim","year":"2018","unstructured":"Lim, L.A., Keles, H.Y.: Foreground segmentation using convolutional neural networks for multiscale feature encoding. Pattern Recognit. Lett. 112, 256\u2013262 (2018)","journal-title":"Pattern Recognit. Lett."},{"issue":"3","key":"5_CR29","doi-asserted-by":"publisher","first-page":"1369","DOI":"10.1007\/s10044-019-00845-9","volume":"23","author":"LA Lim","year":"2019","unstructured":"Lim, L.A., Keles, H.Y.: Learning multi-scale features for foreground segmentation. Pattern Anal. Appl. 23(3), 1369\u20131380 (2019). https:\/\/doi.org\/10.1007\/s10044-019-00845-9","journal-title":"Pattern Anal. Appl."},{"key":"5_CR30","doi-asserted-by":"publisher","first-page":"103584","DOI":"10.1016\/j.cviu.2022.103584","volume":"226","author":"I Kajo","year":"2023","unstructured":"Kajo, I., Kas, M., Ruichek, Y., Kamel, N.: Tensor based completion meets adversarial learning: a win\u2013win solution for change detection on unseen videos. Comput. Vis. Image Underst. 226, 103584 (2023)","journal-title":"Comput. Vis. Image Underst."}],"container-title":["Lecture Notes in Computer Science","Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-49018-7_5","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,26]],"date-time":"2023-11-26T23:09:08Z","timestamp":1701040148000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-49018-7_5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,27]]},"ISBN":["9783031490170","9783031490187"],"references-count":30,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-49018-7_5","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023,11,27]]},"assertion":[{"value":"27 November 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CIARP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Iberoamerican Congress on Pattern Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Coimbra","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Portugal","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 November 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30 November 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ciarp2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Conftool","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"106","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"61","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"58% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}