{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,11]],"date-time":"2024-09-11T07:05:54Z","timestamp":1726038354221},"publisher-location":"Singapore","reference-count":28,"publisher":"Springer Singapore","isbn-type":[{"type":"print","value":"9789811391804"},{"type":"electronic","value":"9789811391811"}],"license":[{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019]]},"DOI":"10.1007\/978-981-13-9181-1_3","type":"book-chapter","created":{"date-parts":[[2019,7,19]],"date-time":"2019-07-19T02:03:22Z","timestamp":1563501802000},"page":"27-40","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["A Novel Foreground Segmentation Method Using Convolutional Neural Network"],"prefix":"10.1007","author":[{"given":"Midhula","family":"Vijayan","sequence":"first","affiliation":[]},{"given":"R.","family":"Mohan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,7,20]]},"reference":[{"key":"3_CR1","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"130","DOI":"10.1007\/978-3-319-25903-1_12","volume-title":"Advanced Concepts for Intelligent Vision Systems","author":"G Allebosch","year":"2015","unstructured":"Allebosch, G., Deboeverie, F., Veelaert, P., Philips, W.: EFIC: edge based foreground background segmentation and interior classification for dynamic camera viewpoints. In: Battiato, S., Blanc-Talon, J., Gallo, G., Philips, W., Popescu, D., Scheunders, P. (eds.) ACIVS 2015. LNCS, vol. 9386, pp. 130\u2013141. Springer, Cham (2015). \n                    https:\/\/doi.org\/10.1007\/978-3-319-25903-1_12"},{"key":"3_CR2","doi-asserted-by":"publisher","first-page":"635","DOI":"10.1016\/j.patcog.2017.09.040","volume":"76","author":"M Babaee","year":"2018","unstructured":"Babaee, M., Dinh, D.T., Rigoll, G.: A deep convolutional neural network for video sequence background subtraction. Pattern Recogn. 76, 635\u2013649 (2018)","journal-title":"Pattern Recogn."},{"issue":"6","key":"3_CR3","doi-asserted-by":"publisher","first-page":"1709","DOI":"10.1109\/TIP.2010.2101613","volume":"20","author":"O Barnich","year":"2011","unstructured":"Barnich, O., Van Droogenbroeck, M.: ViBe: a universal background subtraction algorithm for video sequences. IEEE Trans. Image Process. 20(6), 1709\u20131724 (2011)","journal-title":"IEEE Trans. Image Process."},{"key":"3_CR4","doi-asserted-by":"crossref","unstructured":"Braham, M., Van Droogenbroeck, M.: Deep background subtraction with scene-specific convolutional neural networks. In: 2016 International Conference on Systems, Signals and Image Processing (IWSSIP), pp. 1\u20134. IEEE (2016)","DOI":"10.1109\/IWSSIP.2016.7502717"},{"key":"3_CR5","unstructured":"Chen, Y., Wang, J., Lu, H.: Learning sharable models for robust background subtraction. In: 2015 IEEE International Conference on Multimedia and Expo (ICME), pp. 1\u20136. IEEE (2015)"},{"key":"3_CR6","doi-asserted-by":"publisher","first-page":"35","DOI":"10.1016\/j.cviu.2014.01.004","volume":"122","author":"Z Chen","year":"2014","unstructured":"Chen, Z., Ellis, T.: A self-adaptive Gaussian mixture model. Comput. Vis. Image Underst. 122, 35\u201346 (2014)","journal-title":"Comput. Vis. Image Underst."},{"key":"3_CR7","unstructured":"De Gregorio, M., Giordano, M.: WiSARDrp for change detection in video sequences. In: CVPR 2016, Google Scholar (2016, submitted)"},{"key":"3_CR8","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"751","DOI":"10.1007\/3-540-45053-X_48","volume-title":"Computer Vision\u2014ECCV 2000","author":"A Elgammal","year":"2000","unstructured":"Elgammal, A., Harwood, D., Davis, L.: Non-parametric model for background subtraction. In: Vernon, D. (ed.) ECCV 2000. LNCS, vol. 1843, pp. 751\u2013767. Springer, Heidelberg (2000). \n                    https:\/\/doi.org\/10.1007\/3-540-45053-X_48"},{"key":"3_CR9","doi-asserted-by":"crossref","unstructured":"Hofmann, M., Tiefenbacher, P., Rigoll, G.: Background segmentation with feedback: the pixel-based adaptive segmenter. In: 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 38\u201343. IEEE (2012)","DOI":"10.1109\/CVPRW.2012.6238925"},{"key":"3_CR10","first-page":"163","volume":"70","author":"AB Jagtap","year":"2017","unstructured":"Jagtap, A.B., Hegadi, R.S.: Feature learning for offline handwritten signature verification using convolution neural networks. Int. J. Technol. Hum. Interact. (IJTHI) 70, 163\u2013176 (2017)","journal-title":"Int. J. Technol. Hum. Interact. (IJTHI)"},{"key":"3_CR11","unstructured":"Jain, S.D., Xiong, B., Grauman, K.: FusionSeg: learning to combine motion and appearance for fully automatic segmentation of generic objects in videos. 2(3), 6 (2017). arXiv preprint \n                    arXiv:1701.05384"},{"key":"3_CR12","doi-asserted-by":"publisher","first-page":"2105","DOI":"10.1109\/TCSVT.2017.2711659","volume":"28","author":"S Jiang","year":"2017","unstructured":"Jiang, S., Lu, X.: WeSamBE: a weight-sample-based method for background subtraction. IEEE Trans. Circ. Syst. Video Technol. 28, 2105\u20132115 (2017)","journal-title":"IEEE Trans. Circ. Syst. Video Technol."},{"issue":"3","key":"3_CR13","doi-asserted-by":"publisher","first-page":"172","DOI":"10.1016\/j.rti.2004.12.004","volume":"11","author":"K Kim","year":"2005","unstructured":"Kim, K., Chalidabhongse, T.H., Harwood, D., Davis, L.: Real-time foreground-background segmentation using codebook model. Real-Time Imaging 11(3), 172\u2013185 (2005)","journal-title":"Real-Time Imaging"},{"key":"3_CR14","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1007\/978-3-319-58838-4_6","volume-title":"Pattern Recognition and Image Analysis","author":"I Martins","year":"2017","unstructured":"Martins, I., Carvalho, P., Corte-Real, L., Alba-Castro, J.L.: BMOG: boosted Gaussian mixture model with\u00a0controlled complexity. In: Alexandre, L.A., Salvador S\u00e1nchez, J., Rodrigues, J.M.F. (eds.) IbPRIA 2017. LNCS, vol. 10255, pp. 50\u201357. Springer, Cham (2017). \n                    https:\/\/doi.org\/10.1007\/978-3-319-58838-4_6"},{"issue":"8","key":"3_CR15","doi-asserted-by":"publisher","first-page":"831","DOI":"10.1109\/34.868684","volume":"22","author":"NM Oliver","year":"2000","unstructured":"Oliver, N.M., Rosario, B., Pentland, A.P.: A Bayesian computer vision system for modeling human interactions. IEEE Trans. Pattern Anal. Mach. Intell. 22(8), 831\u2013843 (2000)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"3_CR16","doi-asserted-by":"crossref","unstructured":"Pathak, D., Girshick, R., Doll\u00e1r, P., Darrell, T., Hariharan, B.: Learning features by watching objects move. In: Computer Vision and Pattern Recognition (CVPR) (2017)","DOI":"10.1109\/CVPR.2017.638"},{"key":"3_CR17","doi-asserted-by":"publisher","first-page":"990","DOI":"10.1016\/j.neucom.2015.04.118","volume":"175","author":"G Ram\u00edrez-Alonso","year":"2016","unstructured":"Ram\u00edrez-Alonso, G., Chac\u00f3n-Murgu\u00eda, M.I.: Auto-adaptive parallel som architecture with a modular analysis for dynamic object segmentation in videos. Neurocomputing 175, 990\u20131000 (2016)","journal-title":"Neurocomputing"},{"key":"3_CR18","doi-asserted-by":"crossref","unstructured":"Sajid, H., Cheung, S.C.S.: Background subtraction for static & moving camera. In: 2015 IEEE International Conference on Image Processing (ICIP), pp. 4530\u20134534. IEEE (2015)","DOI":"10.1109\/ICIP.2015.7351664"},{"key":"3_CR19","doi-asserted-by":"crossref","unstructured":"St-Charles, P.L., Bilodeau, G.A., Bergevin, R.: A self-adjusting approach to change detection based on background word consensus. In: 2015 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 990\u2013997. IEEE (2015)","DOI":"10.1109\/WACV.2015.137"},{"issue":"1","key":"3_CR20","doi-asserted-by":"publisher","first-page":"359","DOI":"10.1109\/TIP.2014.2378053","volume":"24","author":"PL St-Charles","year":"2015","unstructured":"St-Charles, P.L., Bilodeau, G.A., Bergevin, R.: SuBSENSE: a universal change detection method with local adaptive sensitivity. IEEE Trans. Image Process. 24(1), 359\u2013373 (2015)","journal-title":"IEEE Trans. Image Process."},{"issue":"10","key":"3_CR21","doi-asserted-by":"publisher","first-page":"4768","DOI":"10.1109\/TIP.2016.2598691","volume":"25","author":"PL St-Charles","year":"2016","unstructured":"St-Charles, P.L., Bilodeau, G.A., Bergevin, R.: Universal background subtraction using word consensus models. IEEE Trans. Image Process. 25(10), 4768\u20134781 (2016)","journal-title":"IEEE Trans. Image Process."},{"key":"3_CR22","unstructured":"Stauffer, C., Grimson, W.E.L.: Adaptive background mixture models for real-time tracking. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 2, pp. 246\u2013252. IEEE (1999)"},{"issue":"1","key":"3_CR23","doi-asserted-by":"publisher","first-page":"25","DOI":"10.1186\/s41074-017-0036-1","volume":"9","author":"A Varghese","year":"2017","unstructured":"Varghese, A., Sreelekha, G.: Sample-based integrated background subtraction and shadow detection. IPSJ Trans. Comput. Vis. Appl. 9(1), 25 (2017)","journal-title":"IPSJ Trans. Comput. Vis. Appl."},{"key":"3_CR24","doi-asserted-by":"publisher","first-page":"963","DOI":"10.1016\/j.ijleo.2018.05.012","volume":"168","author":"M Vijayan","year":"2018","unstructured":"Vijayan, M., Ramasundaram, M.: Moving object detection using vector image model. Optik 168, 963\u2013973 (2018)","journal-title":"Optik"},{"key":"3_CR25","doi-asserted-by":"publisher","first-page":"15505","DOI":"10.1109\/ACCESS.2018.2812880","volume":"6","author":"K Wang","year":"2018","unstructured":"Wang, K., Gou, C., Wang, F.Y.: M4CD: a robust change detection method for intelligent visual surveillance. IEEE Access 6, 15505\u201315520 (2018)","journal-title":"IEEE Access"},{"key":"3_CR26","doi-asserted-by":"crossref","unstructured":"Wang, Y., Jodoin, P.M., Porikli, F., Konrad, J., Benezeth, Y., Ishwar, P.: CDNET 2014: an expanded change detection benchmark dataset. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 387\u2013394 (2014)","DOI":"10.1109\/CVPRW.2014.126"},{"key":"3_CR27","doi-asserted-by":"publisher","first-page":"66","DOI":"10.1016\/j.patrec.2016.09.014","volume":"96","author":"Y Wang","year":"2017","unstructured":"Wang, Y., Luo, Z., Jodoin, P.M.: Interactive deep learning method for segmenting moving objects. Pattern Recogn. Lett. 96, 66\u201375 (2017)","journal-title":"Pattern Recogn. Lett."},{"issue":"1","key":"3_CR28","doi-asserted-by":"publisher","first-page":"254","DOI":"10.1109\/TITS.2017.2754099","volume":"19","author":"L Yang","year":"2018","unstructured":"Yang, L., Li, J., Luo, Y., Zhao, Y., Cheng, H., Li, J.: Deep background modeling using fully convolutional network. IEEE Trans. Intell. Transp. Syst. 19(1), 254\u2013262 (2018)","journal-title":"IEEE Trans. Intell. Transp. Syst."}],"container-title":["Communications in Computer and Information Science","Recent Trends in Image Processing and Pattern Recognition"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-13-9181-1_3","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2019,7,19]],"date-time":"2019-07-19T02:04:04Z","timestamp":1563501844000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-981-13-9181-1_3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019]]},"ISBN":["9789811391804","9789811391811"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-981-13-9181-1_3","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2019]]},"assertion":[{"value":"20 July 2019","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"RTIP2R","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Recent Trends in Image Processing and Pattern Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Solapur","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"India","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2018","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21 December 2018","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 December 2018","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"rtip2018","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/rtip2r.org\/2018\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"374","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"173","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"46% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"-","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}