{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T00:54:19Z","timestamp":1740099259984,"version":"3.37.3"},"publisher-location":"Cham","reference-count":28,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030057152"},{"type":"electronic","value":"9783030057169"}],"license":[{"start":{"date-parts":[[2018,12,11]],"date-time":"2018-12-11T00:00:00Z","timestamp":1544486400000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019]]},"DOI":"10.1007\/978-3-030-05716-9_41","type":"book-chapter","created":{"date-parts":[[2018,12,10]],"date-time":"2018-12-10T14:17:32Z","timestamp":1544451452000},"page":"506-518","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Utilizing Deep Object Detector for Video Surveillance Indexing and Retrieval"],"prefix":"10.1007","author":[{"given":"Tom","family":"Durand","sequence":"first","affiliation":[]},{"given":"Xiyan","family":"He","sequence":"additional","affiliation":[]},{"given":"Ionel","family":"Pop","sequence":"additional","affiliation":[]},{"given":"Lionel","family":"Robinault","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2018,12,11]]},"reference":[{"issue":"3","key":"41_CR1","doi-asserted-by":"crossref","first-page":"187","DOI":"10.3169\/mta.4.187","volume":"4","author":"G Awad","year":"2016","unstructured":"Awad, G., Snoek, C.G.M., Smeaton, A.F., Qu\u00e9not, G.: Trecvid semantic indexing of video: a 6-year retrospective. ITE Trans. Media Technol. Appl. 4(3), 187\u2013208 (2016)","journal-title":"ITE Trans. Media Technol. Appl."},{"issue":"2","key":"41_CR2","doi-asserted-by":"crossref","first-page":"295","DOI":"10.1109\/TPAMI.2015.2439281","volume":"38","author":"C Dong","year":"2016","unstructured":"Dong, C., Loy, C.C., He, K., Tang, X.: Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 38(2), 295\u2013307 (2016)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"41_CR3","unstructured":"Fu, C.Y., Liu, W., Ranga, A., Tyagi, A., Berg, A.C.: DSSD : deconvolutional single shot detector. arXiv preprint arXiv:1701.06659 (2017)"},{"key":"41_CR4","series-title":"Advances in Intelligent Systems and Computing","doi-asserted-by":"publisher","first-page":"73","DOI":"10.1007\/978-3-319-23814-2_9","volume-title":"Image Processing and Communications Challenges 7","author":"M Fularz","year":"2016","unstructured":"Fularz, M., Kraft, M., Schmidt, A., Niechcia\u0142, J.: The PUT surveillance database. In: Chora\u015b, R.S. (ed.) Image Processing and Communications Challenges 7. AISC, vol. 389, pp. 73\u201379. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-23814-2_9"},{"key":"41_CR5","doi-asserted-by":"crossref","unstructured":"Girshick, R.B.: Fast r-cnn. In: ICCV, pp. 1440\u20131448. IEEE Press, Santiago (2015)","DOI":"10.1109\/ICCV.2015.169"},{"key":"41_CR6","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., Girshick, R.B.: Mask R-CNN. In: ICCV, pp. 2980\u20132988. IEEE Press, Venise (2017)","DOI":"10.1109\/ICCV.2017.322"},{"key":"41_CR7","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR, pp. 770\u2013778. IEEE Press, Las Vegas (2016)","DOI":"10.1109\/CVPR.2016.90"},{"issue":"6","key":"41_CR8","doi-asserted-by":"crossref","first-page":"797","DOI":"10.1109\/TSMCC.2011.2109710","volume":"41","author":"W Hu","year":"2011","unstructured":"Hu, W., Xie, N., Li, L., Zeng, X., Maybank, S.: A survey on visual content-based video indexing and retrieval. IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.) 41(6), 797\u2013819 (2011)","journal-title":"IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.)"},{"issue":"3","key":"41_CR9","doi-asserted-by":"crossref","first-page":"334","DOI":"10.1109\/TSMCC.2004.829274","volume":"34","author":"W Hu","year":"2004","unstructured":"Hu, W., Tan, T., Wang, L., Maybank, S.: A survey on visual surveillance of object motion and behaviors. IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.) 34(3), 334\u2013352 (2004)","journal-title":"IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.)"},{"key":"41_CR10","unstructured":"Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: ICML, pp. 448\u2013456. JMLR.org (2015)"},{"key":"41_CR11","doi-asserted-by":"crossref","unstructured":"Jung, H., Choi, M.K., Jung, J., Lee, J.H., Kwon, S., Jung, W.Y.: Resnet-based vehicle classification and localization in traffic surveillance systems. In: CVPRW, pp. 934\u2013940. IEEE Press, Honolulu (2017)","DOI":"10.1109\/CVPRW.2017.129"},{"key":"41_CR12","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems, vol. 25, pp. 1097\u20131105. Curran Associates Inc., Lake Tahoe (2012)"},{"key":"41_CR13","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Goyal, P., Girshick, R.B., He, K., Doll\u00e1r, P.: Focal loss for dense object detection. In: ICCV, pp. 2999\u20133007. IEEE Press, Venise (2017)","DOI":"10.1109\/ICCV.2017.324"},{"key":"41_CR14","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"21","DOI":"10.1007\/978-3-319-46448-0_2","volume-title":"Computer Vision \u2013 ECCV 2016","author":"W Liu","year":"2016","unstructured":"Liu, W., et al.: SSD: single shot multibox detector. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 21\u201337. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46448-0_2"},{"key":"41_CR15","doi-asserted-by":"crossref","unstructured":"Luo, Z., et al.: MIO-TCD: a new benchmark dataset for vehicle classification and localization. IEEE Trans. Image Process. 27, 5129\u20135141 (2018)","DOI":"10.1109\/TIP.2018.2848705"},{"key":"41_CR16","doi-asserted-by":"crossref","unstructured":"Ning, G., et al.: Spatially supervised recurrent convolutional neural networks for visual object tracking. In: ISCAS, pp. 1\u20134. IEEE Press, Baltimore (2017)","DOI":"10.1109\/ISCAS.2017.8050867"},{"key":"41_CR17","series-title":"Lecture Notes in Networks and Systems","doi-asserted-by":"publisher","first-page":"359","DOI":"10.1007\/978-3-319-56991-8_27","volume-title":"Proceedings of SAI Intelligent Systems Conference (IntelliSys) 2016","author":"A Podlesnaya","year":"2018","unstructured":"Podlesnaya, A., Podlesnyy, S.: Deep learning based semantic video indexing and retrieval. In: Bi, Y., Kapoor, S., Bhatia, R. (eds.) IntelliSys 2016. LNNS, vol. 16, pp. 359\u2013372. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-319-56991-8_27"},{"key":"41_CR18","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S.K., Girshick, R.B., Farhadi, A.: You only look once: unified, real-time object detection. In: CVPR, pp. 779\u2013788. IEEE Press, Las Vegas (2016)","DOI":"10.1109\/CVPR.2016.91"},{"key":"41_CR19","doi-asserted-by":"crossref","unstructured":"Redmon, J., Farhadi, A.: Yolo9000: better, faster, stronger. In: CVPR, pp. 6517\u20136525. IEEE Press, Honolulu (2017)","DOI":"10.1109\/CVPR.2017.690"},{"key":"41_CR20","unstructured":"Redmon, J., Farhadi, A.: Yolov3: an incremental improvement. arXiv preprint arXiv:1804.02767 (2018)"},{"issue":"6","key":"41_CR21","doi-asserted-by":"crossref","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","volume":"39","author":"S Ren","year":"2015","unstructured":"Ren, S., He, K., Girshick, R.B., Sun, J.B.: Faster r-cnn: towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(6), 1137\u20131149 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"3","key":"41_CR22","doi-asserted-by":"crossref","first-page":"222","DOI":"10.1007\/s11263-013-0636-x","volume":"105","author":"J S\u00e1nchez","year":"2013","unstructured":"S\u00e1nchez, J., Perronnin, F., Mensink, T., Verbeek, J.J.: Image classification with the fisher vector: theory and practice. Int. J. Comput. Vis. 105(3), 222\u2013245 (2013)","journal-title":"Int. J. Comput. Vis."},{"key":"41_CR23","doi-asserted-by":"crossref","unstructured":"Stauffer, C., Grimson, W.E.L.: Adaptive background mixture models for real-time tracking. In: CVPR, pp. 2246\u20132252. IEEE Press, Ft. Collins (1999)","DOI":"10.1109\/CVPR.1999.784637"},{"key":"41_CR24","doi-asserted-by":"crossref","unstructured":"Szegedy, C., et al.: Going deeper with convolutions. In: CVPR, pp. 1\u20139. IEEE Press, Boston (2015)","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"41_CR25","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In: CVPR, pp. 2818\u20132826. IEEE Press, Las Vegas (2016)","DOI":"10.1109\/CVPR.2016.308"},{"key":"41_CR26","doi-asserted-by":"crossref","unstructured":"Ueki, K., Kobayashi, T.: Object detection oriented feature pooling for video semantic indexing. In: VISIGRAPP, pp. 44\u201351. SciTePress (2017)","DOI":"10.5220\/0006099600440051"},{"key":"41_CR27","doi-asserted-by":"crossref","unstructured":"Wang, Z., Chang, S., Yang, Y., Liu, D., Huang, T.S.: Studying very low resolution recognition using deep networks. In: CVPR, pp. 4792\u20134800. IEEE Press, Las Vegas (2016)","DOI":"10.1109\/CVPR.2016.518"},{"key":"41_CR28","doi-asserted-by":"crossref","unstructured":"Xu, Z., Hu, J., Deng, W.: Recurrent convolutional neural network for video classification. In: ICME, pp. 1\u20136. IEEE Press, Seattle (2016)","DOI":"10.1109\/ICME.2016.7552971"}],"container-title":["Lecture Notes in Computer Science","MultiMedia Modeling"],"original-title":[],"link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-05716-9_41","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2019,11,6]],"date-time":"2019-11-06T23:52:40Z","timestamp":1573084360000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-05716-9_41"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,12,11]]},"ISBN":["9783030057152","9783030057169"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-05716-9_41","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2018,12,11]]},"assertion":[{"value":"MMM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Modeling","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Thessaloniki","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Greece","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2019","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 January 2019","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"11 January 2019","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmm2019","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/mmm2019.iti.gr\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double blind for full papers and workshop papers, single blind for other paper types","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"204","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"96","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"47% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"2.67","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"6 demonstration papers, 5 industry papers, 6 workshop papers, and 6 Video Browser Showdown papers were also accepted.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}}]}}