{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,1]],"date-time":"2025-10-01T15:35:35Z","timestamp":1759332935365,"version":"3.40.3"},"publisher-location":"Cham","reference-count":26,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030164461"},{"type":"electronic","value":"9783030164478"}],"license":[{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019]]},"DOI":"10.1007\/978-3-030-16447-8_9","type":"book-chapter","created":{"date-parts":[[2019,3,29]],"date-time":"2019-03-29T23:04:19Z","timestamp":1553900659000},"page":"85-94","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["A Machine Learning Approach to Detect Violent Behaviour from Video"],"prefix":"10.1007","author":[{"given":"David","family":"Nova","sequence":"first","affiliation":[]},{"given":"Andr\u00e9","family":"Ferreira","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7991-2090","authenticated-orcid":false,"given":"Paulo","family":"Cortez","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,3,31]]},"reference":[{"issue":"20","key":"9_CR1","doi-asserted-by":"publisher","first-page":"6935","DOI":"10.1016\/j.eswa.2015.05.023","volume":"42","author":"P Afsar","year":"2015","unstructured":"Afsar, P., Cortez, P., Santos, H.: Automatic visual detection of human behavior: a review from 2000 to 2014. Expert Syst. Appl. 42(20), 6935\u20136956 (2015). \n                    https:\/\/doi.org\/10.1016\/j.eswa.2015.05.023","journal-title":"Expert Syst. Appl."},{"key":"9_CR2","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1016\/j.eswa.2018.03.035","volume":"110","author":"P Afsar","year":"2018","unstructured":"Afsar, P., Cortez, P., Santos, H.M.D.: Automatic human trajectory destination prediction from video. Expert Syst. Appl. 110, 41\u201351 (2018). \n                    https:\/\/doi.org\/10.1016\/j.eswa.2018.03.035","journal-title":"Expert Syst. Appl."},{"key":"9_CR3","doi-asserted-by":"crossref","unstructured":"Cao, Z., Simon, T., Wei, S.E., Sheikh, Y.: Realtime multi-person 2D pose estimation using part affinity fields. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.143"},{"key":"9_CR4","unstructured":"Ch\u00e9ron, G., Laptev, I., Schmid, C.: P-CNN: pose-based CNN features for action recognition. CoRR abs\/1506.03607 (2015). \n                    http:\/\/arxiv.org\/abs\/1506.03607"},{"key":"9_CR5","unstructured":"Clarin, C.T., Dionisio, J.A.M., Echavez, M.T., Naval, P.C.: DOVE: detection of movie violence using motion intensity analysis on skin and blood. Technical report, University of the Philippines (2005)"},{"key":"9_CR6","doi-asserted-by":"crossref","unstructured":"Coppola, C., Faria, D., Nunes, U., Bellotto, N.: Social activity recognition based on probabilistic merging of skeleton features with proximity priors from RGB-D data. In: Proceedings of IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5055\u20135061 (2016)","DOI":"10.1109\/IROS.2016.7759742"},{"key":"9_CR7","doi-asserted-by":"publisher","unstructured":"Datta, A., Shah, M., Lobo, N.D.V.: Person-on-person violence detection in video data. In: Object Recognition Supported by User Interaction for Service Robots, vol. 1, pp. 433\u2013438, August 2002. \n                    https:\/\/doi.org\/10.1109\/ICPR.2002.1044748","DOI":"10.1109\/ICPR.2002.1044748"},{"key":"9_CR8","unstructured":"Deniz, O., Serrano, I., Bueno, G., Kim, T.: Fast violence detection in video. In: 2014 International Conference on Computer Vision Theory and Applications (VISAPP), vol. 2, pp. 478\u2013485, January 2014"},{"key":"9_CR9","doi-asserted-by":"publisher","first-page":"517","DOI":"10.1007\/978-981-10-3002-4_43","volume-title":"Pattern Recognition","author":"Z Dong","year":"2016","unstructured":"Dong, Z., Qin, J., Wang, Y.: Multi-stream deep networks for person to person violence detection in videos. In: Tan, T., Li, X., Chen, X., Zhou, J., Yang, J., Cheng, H. (eds.) Pattern Recognition, pp. 517\u2013531. Springer, Singapore (2016). \n                    https:\/\/doi.org\/10.1007\/978-981-10-3002-4_43"},{"key":"9_CR10","doi-asserted-by":"publisher","unstructured":"Du, W., Wang, Y., Qiao, Y.: RPAN: an end-to-end recurrent pose-attention network for action recognition in videos. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 3745\u20133754, October 2017. \n                    https:\/\/doi.org\/10.1109\/ICCV.2017.402","DOI":"10.1109\/ICCV.2017.402"},{"key":"9_CR11","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. CoRR abs\/1512.03385 (2015). \n                    http:\/\/arxiv.org\/abs\/1512.03385"},{"key":"9_CR12","unstructured":"Henriques, J.F., Caseiro, R., Martins, P., Batista, J.: High-speed tracking with kernelized correlation filters. CoRR abs\/1404.7584 (2014). \n                    http:\/\/arxiv.org\/abs\/1404.7584"},{"key":"9_CR13","unstructured":"Herath, S., Harandi, M.T., Porikli, F.: Going deeper into action recognition: a survey. CoRR abs\/1605.04988 (2016). \n                    http:\/\/arxiv.org\/abs\/1605.04988"},{"key":"9_CR14","unstructured":"Kong, Y., Fu, Y.: Human Action Recognition and Prediction: A Survey. ArXiv e-prints, June 2018"},{"key":"9_CR15","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: Pereira, F., Burges, C.J.C., Bottou, L., Weinberger, K.Q. (eds.) Advances in Neural Information Processing Systems, vol. 25, pp. 1097\u20131105. Curran Associates, Inc. (2012). \n                    http:\/\/papers.nips.cc\/paper\/4824-imagenet-classification-with-deep-convolutional-neural-networks.pdf"},{"key":"9_CR16","unstructured":"Li, L., Zheng, W., Zhang, Z., Huang, Y., Wang, L.: Skeleton-based relational modeling for action recognition. CoRR abs\/1805.02556 (2018). \n                    http:\/\/arxiv.org\/abs\/1805.02556"},{"key":"9_CR17","unstructured":"Liu, J., Shahroudy, A., Xu, D., Wang, G.: Spatio-temporal LSTM with trust gates for 3D human action recognition. CoRR abs\/1607.07043 (2016). \n                    http:\/\/arxiv.org\/abs\/1607.07043"},{"key":"9_CR18","unstructured":"Luvizon, D.C., Picard, D., Tabia, H.: 2D\/3D pose estimation and action recognition using multitask deep learning. CoRR abs\/1802.09232 (2018). \n                    http:\/\/arxiv.org\/abs\/1802.09232"},{"key":"9_CR19","unstructured":"Ng, A.: Machine Learning Yearning. deeplearning.ai (2018)"},{"key":"9_CR20","unstructured":"Pham, H., Khoudour, L., Crouzil, A., Zegers, P., Velastin, S.A.: Exploiting deep residual networks for human action recognition from skeletal data. CoRR abs\/1803.07781 (2018). \n                    http:\/\/arxiv.org\/abs\/1803.07781"},{"key":"9_CR21","unstructured":"Pham, H., Khoudour, L., Crouzil, A., Zegers, P., Velastin, S.A.: Learning and recognizing human action from skeleton movement with deep residual neural networks. CoRR abs\/1803.07780 (2018). \n                    http:\/\/arxiv.org\/abs\/1803.07780"},{"key":"9_CR22","unstructured":"Sudhakaran, S., Lanz, O.: Learning to detect violent videos using convolutional long short-term memory. CoRR abs\/1709.06531 (2017). \n                    http:\/\/arxiv.org\/abs\/1709.06531"},{"key":"9_CR23","doi-asserted-by":"publisher","unstructured":"Vasconcelos, N., Lippman, A.: Towards semantically meaningful feature spaces for the characterization of video content. In: Proceedings of International Conference on Image Processing, vol. 1, pp. 25\u201328, October 1997. \n                    https:\/\/doi.org\/10.1109\/ICIP.1997.647375","DOI":"10.1109\/ICIP.1997.647375"},{"key":"9_CR24","unstructured":"Wang, Q.: A survey of visual analysis of human motion and its applications. CoRR abs\/1608.00700 (2016). \n                    http:\/\/arxiv.org\/abs\/1608.00700"},{"key":"9_CR25","volume-title":"Data Mining: Practical Machine Learning Tools and Techniques","author":"I Witten","year":"2017","unstructured":"Witten, I., Frank, E., Hall, M., Pal, C.: Data Mining: Practical Machine Learning Tools and Techniques, 4th edn. Morgan Kaufmann, San Franscico (2017)","edition":"4"},{"key":"9_CR26","unstructured":"Zolfaghari, M., Oliveira, G.L., Sedaghat, N., Brox, T.: Chained multi-stream networks exploiting pose, motion, and appearance for action classification and detection. CoRR abs\/1704.00616 (2017). \n                    http:\/\/arxiv.org\/abs\/1704.00616"}],"container-title":["Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering","Intelligent Technologies for Interactive Entertainment"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-16447-8_9","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2019,5,21]],"date-time":"2019-05-21T01:40:40Z","timestamp":1558402840000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-16447-8_9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019]]},"ISBN":["9783030164461","9783030164478"],"references-count":26,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-16447-8_9","relation":{},"ISSN":["1867-8211","1867-822X"],"issn-type":[{"type":"print","value":"1867-8211"},{"type":"electronic","value":"1867-822X"}],"subject":[],"published":{"date-parts":[[2019]]},"assertion":[{"value":"31 March 2019","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"INTETAIN","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Intelligent Technologies for Interactive Entertainment","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Guimar\u00e3es","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Portugal","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2018","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21 November 2018","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 November 2018","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"10","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"intetain2018","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.allconferencealert.org\/intetain-2018\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Open","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"confy.eai.eu","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"23","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"15","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"65% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}}]}}