{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T16:41:34Z","timestamp":1772642494796,"version":"3.50.1"},"reference-count":34,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2020,3,18]],"date-time":"2020-03-18T00:00:00Z","timestamp":1584489600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2020,3,18]],"date-time":"2020-03-18T00:00:00Z","timestamp":1584489600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Big Data"],"published-print":{"date-parts":[[2020,12]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Multi activity-multi object recognition (MAMO) is a challenging task in visual systems for monitoring, recognizing and alerting in various public places, such as universities, hospitals and airports. While both academic and commercial researchers are aiming towards automatic tracking of human activities in intelligent video surveillance using deep learning frameworks. This is required for many real time applications to detect unusual\/suspicious activities like tracking of suspicious behaviour in crime events etc. The primary purpose of this paper is to render a multi class activity prediction in individuals as well as groups from video sequences by using the state-of-the-art object detector You Look only Once (YOLOv3). By optimum utilization of the geographical information of cameras and YOLO object detection framework, a Deep Landmark model recognize a simple to complex human actions on gray scale to RGB image frames of video sequences. This model is tested and compared with various benchmark datasets and found to be the most precise model for detecting human activities in video streams. Upon analysing the experimental results, it has been observed that the proposed method shows superior performance as well as high accuracy.<\/jats:p>","DOI":"10.1186\/s40537-020-00296-8","type":"journal-article","created":{"date-parts":[[2020,3,18]],"date-time":"2020-03-18T14:02:46Z","timestamp":1584540166000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["A comparison on visual prediction models for MAMO (multi activity-multi object) recognition using deep learning"],"prefix":"10.1186","volume":"7","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-5327-9795","authenticated-orcid":false,"given":"Budi","family":"Padmaja","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4734-5914","authenticated-orcid":false,"given":"Madhu Bala","family":"Myneni","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7846-1286","authenticated-orcid":false,"given":"Epili","family":"Krishna Rao Patro","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,3,18]]},"reference":[{"key":"296_CR1","doi-asserted-by":"crossref","unstructured":"Singh S, Velastin SA, Ragheb H. Muhavi: a multicamera human action video dataset for the evaluation of action recognition methods. In: 2010 seventh IEEE international conference on advanced video and signal based surveillance (AVSS); 2010. p. 48\u201355.","DOI":"10.1109\/AVSS.2010.63"},{"issue":"5","key":"296_CR2","doi-asserted-by":"publisher","first-page":"893","DOI":"10.1109\/TPAMI.2007.70731","volume":"30","author":"T Xiang","year":"2008","unstructured":"Xiang T, Gong S. Video behavior profiling for anomaly detection. IEEE Trans Pattern Anal Mach Intell. 2008;30(5):893\u2013908.","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"296_CR3","unstructured":"Shirai A, Geslin E, Richir S. Wiimedia: motion analysis methods and applications using a consumer video game controller. In: Proceedings of the 2007 ACM SIGGRAPH symposium on video games, 2007. New York: ACM; 2007. p. 133\u201340."},{"key":"296_CR4","doi-asserted-by":"publisher","first-page":"47","DOI":"10.1007\/978-3-540-72348-6_3","volume-title":"Artifical intelligence for human computing","author":"M Pantic","year":"2007","unstructured":"Pantic M, Pentland A, Nijholt A, Huang T. Human computing and machine understanding of human behavior: a survey. Artifical intelligence for human computing. Berlin: Springer; 2007. p. 47\u201371."},{"key":"296_CR5","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1007\/10705432_17","volume-title":"Cooperative buildings: integrating information, organizations, and architecture","author":"C Kidd","year":"1999","unstructured":"Kidd C, Orr R, Abowd G, Atkeson C, Essa I, MacIntyre B, Mynatt E, Starner T, Newstetter W. The aware home: a living laboratory for ubiquitous computing research. Cooperative buildings: integrating information, organizations, and architecture. Berlin: Springer; 1999. p. 191\u20138."},{"key":"296_CR6","doi-asserted-by":"publisher","first-page":"976","DOI":"10.1016\/j.imavis.2009.11.014","volume":"28","author":"R Poppe","year":"2010","unstructured":"Poppe R. A survey on vision-based human action recognition. Image Vis Comput. 2010;28:976\u201390.","journal-title":"Image Vis Comput"},{"key":"296_CR7","doi-asserted-by":"publisher","first-page":"22","DOI":"10.1007\/s11263-010-0384-0","volume":"93","author":"Q Shi","year":"2011","unstructured":"Shi Q, Cheng L, Wang L, Smola A. Human action segmentation and recognition using discriminative semi-markov models. IJCV. 2011;93:22\u201332.","journal-title":"IJCV"},{"key":"296_CR8","doi-asserted-by":"crossref","unstructured":"Feichtenhofer C, Pinz A, Wildes RP. Spatiotemporal multiplier networks for video action recognition. In: IEEE conference on computer vision and pattern recognition (CVPR); 2017. p. 7445\u201354.","DOI":"10.1109\/CVPR.2017.787"},{"issue":"3","key":"296_CR9","doi-asserted-by":"publisher","first-page":"257","DOI":"10.1109\/34.910878","volume":"23","author":"A Bobick","year":"2001","unstructured":"Bobick A, Davis J. The recognition of human movement using temporal templates. IEEE Trans Pattern Anal Mach Intell. 2001;23(3):257\u201367.","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"2\u20133","key":"296_CR10","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1016\/j.cviu.2006.07.013","volume":"104","author":"D Weinland","year":"2006","unstructured":"Weinland D, Ronfard R, Boyer E. Free viewpoint action recognition using motion history volumes. Comput Vis Image Underst. 2006;104(2\u20133):249\u201357.","journal-title":"Comput Vis Image Underst"},{"key":"296_CR11","doi-asserted-by":"crossref","unstructured":"Klaser A, Marszalek M, Schmid C. A spatio-temporal descriptor based on 3D-gradients. In: BMVC, 2008; 2008.","DOI":"10.5244\/C.22.99"},{"key":"296_CR12","doi-asserted-by":"crossref","unstructured":"Wang H, Kl\u00e4ser A, Schmid C, Liu CL. Action recognition by dense trajectories. In: IEEE conference on computer vision & pattern recognition, Colorado Springs, United States; 2011. p. 3169\u201376.","DOI":"10.1109\/CVPR.2011.5995407"},{"key":"296_CR13","doi-asserted-by":"crossref","unstructured":"Bregonzio M, Gong S, Xiang T. Recognizing action as clouds of space-time interest points. In: CVPR, 2009; 2009.","DOI":"10.1109\/CVPRW.2009.5206779"},{"key":"296_CR14","doi-asserted-by":"crossref","unstructured":"Laptev I, Marszalek M, Schmid C, Rozenfeld B. Learning realistic human actions from movies. In: CVPR, 2008; 2008.","DOI":"10.1109\/CVPR.2008.4587756"},{"issue":"5","key":"296_CR15","doi-asserted-by":"publisher","first-page":"1813","DOI":"10.1016\/j.patcog.2013.12.004","volume":"47","author":"J Cho","year":"2014","unstructured":"Cho J, Lee M, Chang HJ, Oh S. Robust action recognition using local motion and group sparsity. Pattern Recognit. 2014;47(5):1813\u201325.","journal-title":"Pattern Recognit"},{"key":"296_CR16","doi-asserted-by":"crossref","unstructured":"Padmaja B, Rao PN, Bala MM, Patro EKR. A novel design of autonomous cars using IoT and visual features. In: 2018 2nd international conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud) (I-SMAC) I-SMAC (IoT in Social, Mobile, Analytics and Cloud) (I-SMAC), 2018. New York: IEEE; 2018. p. 18\u201321.","DOI":"10.1109\/I-SMAC.2018.8653736"},{"key":"296_CR17","unstructured":"Ryoo M, Aggarwal J. Recognition of composite human activities through context-free grammar based representation. In: CVPR, vol 2; 2006. p. 1709\u201318."},{"key":"296_CR18","doi-asserted-by":"publisher","first-page":"215","DOI":"10.1007\/978-3-642-33765-9_16","volume-title":"Computer Vision \u2013 ECCV 2012","author":"Wongun Choi","year":"2012","unstructured":"Choi W, Savarese S. A unified framework for multi-target tracking and collective activity recognition. In: ECCV. Berlin: Springer; 2012. p. 215\u201330."},{"key":"296_CR19","doi-asserted-by":"crossref","unstructured":"Hadfield S, Bowden R. Hollywood 3D: recognizing actions in 3D natural scenes. In: CVPR, Portland, Oregon; 2013.","DOI":"10.1109\/CVPR.2013.436"},{"key":"296_CR20","doi-asserted-by":"crossref","unstructured":"Donahue J, Hendricks L, Guadarrama S, Rohrbach M, Venugopalan S, Saenko K, Darrell T. Long-term recurrent convolutional networks for visual recognition and description. In: CVPR, 2015; 2015.","DOI":"10.1109\/CVPR.2015.7298878"},{"key":"296_CR21","doi-asserted-by":"crossref","unstructured":"Taylor GW, Fergus R, LeCun Y, Bregler C. Convolutional learning of spatio-temporal features. In: ECCV, 2010; 2010.","DOI":"10.1007\/978-3-642-15567-3_11"},{"key":"296_CR22","unstructured":"Simonyan K, Zisserman A. Two-stream convolutional networks for action recognition in videos. In: NIPS, 2014; 2014."},{"key":"296_CR23","unstructured":"Ng JYH, Hausknecht M, Vijayanarasimhan S, Vinyals O, Monga R, Toderici G. Beyond short snippets: deep networks for video classification. In: CVPR, 2015; 2015."},{"issue":"5S","key":"296_CR24","first-page":"1102","volume":"10","author":"B Padmaja","year":"2018","unstructured":"Padmaja B, Rama Prasad VV, Sunitha KVN, Vineeth Reddy G. Deep RNN based human activity recognition using LSTM architecture on smart phone sensor data. J Fundam Appl Sci. 2018;10(5S):1102\u201315.","journal-title":"J Fundam Appl Sci"},{"key":"296_CR25","doi-asserted-by":"crossref","unstructured":"Yan W, Gao Y, Liu Q. Human-object interaction recognition using multitask neural network. In: 2019 3rd international symposium on autonomous systems (ISAS), Shanghai, China; 2019. p. 323\u20138.","DOI":"10.1109\/ISASS.2019.8757767"},{"key":"296_CR26","doi-asserted-by":"crossref","unstructured":"Liu J, Luo J, Shah M. Recognizing realistic actions from videos in the wild. In: CVPR 2009, Miami, FL; 2009.","DOI":"10.1109\/CVPR.2009.5206744"},{"key":"296_CR27","doi-asserted-by":"crossref","unstructured":"Jhuang H, Serre T, Wolf L, Poggio T. A biologically inspired system for action recognition. In: IEEE 11th international conference on computer vision; 2007. p. 1\u20138.","DOI":"10.1109\/ICCV.2007.4408988"},{"key":"296_CR28","unstructured":"Lin Z, Jiang Z, Davis LS. Recognizing actions by shape-motion prototype trees. In: IEEE 12th international conference on computer vision; 2009. p. 444\u201351."},{"key":"296_CR29","doi-asserted-by":"crossref","unstructured":"Liu J, Luo J, Shah M. Recognizing realistic actions from videos in the wild. In: IEEE conference on computer vision and pattern recognition; 2009. p. 1996\u20132003.","DOI":"10.1109\/CVPR.2009.5206744"},{"key":"296_CR30","doi-asserted-by":"crossref","unstructured":"Kim TK, Wong SF, Cipolla R. Tensor canonical correlation analysis for action classification. In: IEEE conference on computer vision and pattern recognition; 2007. p. 1\u20138.","DOI":"10.1109\/CVPR.2007.383137"},{"key":"296_CR31","unstructured":"Ravanbakhsh M, Mousavi H, Mohammad R, Murino V, Davis LS. Action recognition with image based CNN features. In: IEEE conference on computer vision and pattern recognition (CVPR), December 2015; 2015."},{"key":"296_CR32","doi-asserted-by":"crossref","unstructured":"Ulutan O, Swati R, Srivatsa M, Torres C, Manjunath BS. Actor conditioned attention maps for video action detection. Computer Vision and Pattern Recognition; 2019.","DOI":"10.1109\/WACV45572.2020.9093617"},{"key":"296_CR33","doi-asserted-by":"publisher","first-page":"1242","DOI":"10.1109\/TPAMI.2013.220","volume":"36","author":"W Choi","year":"2014","unstructured":"Choi W, Savarese S. Understanding collective activities of people from videos. IEEE Trans Pattern Anal Mach Intell. 2014;36:1242\u201357.","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"296_CR34","doi-asserted-by":"crossref","unstructured":"Choi W, Shahid K, Savarese S. Learning context for collective activity recognition. In: IEEE conference on computer vision and pattern recognition (CVPR); 2011.","DOI":"10.1109\/CVPR.2011.5995707"}],"container-title":["Journal of Big Data"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1186\/s40537-020-00296-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/article\/10.1186\/s40537-020-00296-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1186\/s40537-020-00296-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,3,18]],"date-time":"2021-03-18T00:06:54Z","timestamp":1616026014000},"score":1,"resource":{"primary":{"URL":"https:\/\/journalofbigdata.springeropen.com\/articles\/10.1186\/s40537-020-00296-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,3,18]]},"references-count":34,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2020,12]]}},"alternative-id":["296"],"URL":"https:\/\/doi.org\/10.1186\/s40537-020-00296-8","relation":{},"ISSN":["2196-1115"],"issn-type":[{"value":"2196-1115","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,3,18]]},"assertion":[{"value":"3 September 2019","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 February 2020","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 March 2020","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"The authors declare that they have no competing interests.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"24"}}