{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,19]],"date-time":"2026-02-19T02:57:03Z","timestamp":1771469823091,"version":"3.50.1"},"reference-count":28,"publisher":"Association for Computing Machinery (ACM)","issue":"11","license":[{"start":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T00:00:00Z","timestamp":1732147200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Asian Low-Resour. Lang. Inf. Process."],"published-print":{"date-parts":[[2024,11,30]]},"abstract":"<jats:p>The advancement of medicine presents challenges for modern cultures, especially with unpredictable elderly falling incidents anywhere due to serious health issues. Delayed rescue for at-risk elders can be dangerous. Traditional elder safety methods like video surveillance or wearable sensors are inefficient and burdensome, wasting human resources and requiring caregivers' constant fall detection monitoring. Thus, a more effective and convenient solution is needed to ensure elderly safety. In this article, a method is presented for detecting human falls in naturally occurring scenes using videos through a traditional Convolutional Neural Network (CNN) model, Inception-v3, VGG-19, and two versions of the You Only Look Once (YOLO) working model. The primary focus of this work is human fall detection through the utilization of deep learning models. Specifically, the YOLO approach is adopted for object detection and tracking in video scenes. By implementing YOLO, human subjects are identified, and bounding boxes are generated around them. The classification of various human activities, including fall detection is accomplished through the analysis of deformation features extracted from these bounding boxes. The traditional CNN model achieves an impressive 99.83% accuracy in human fall detection, surpassing other state-of-the-art methods. However, training time is longer compared to YOLO-v2 and YOLO-v3, but significantly shorter than Inception-v3, taking only around 10% of its total training time.<\/jats:p>","DOI":"10.1145\/3687125","type":"journal-article","created":{"date-parts":[[2024,8,10]],"date-time":"2024-08-10T11:23:02Z","timestamp":1723288982000},"page":"1-16","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":3,"title":["Learning and Vision-based approach for Human fall detection and classification in naturally occurring scenes using video data"],"prefix":"10.1145","volume":"23","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-6974-7830","authenticated-orcid":false,"given":"Shashvat","family":"Singh","sequence":"first","affiliation":[{"name":"Computer Science, Banaras Hindu University, Varanasi, India"}]},{"ORCID":"https:\/\/orcid.org\/0009-0001-2381-5531","authenticated-orcid":false,"given":"Kumkum","family":"Kumari","sequence":"additional","affiliation":[{"name":"Computer Science, Banaras Hindu University, Varanasi, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2479-1153","authenticated-orcid":false,"given":"Ankita","family":"Vaish","sequence":"additional","affiliation":[{"name":"Computer Science, Banaras Hindu University Faculty of Science, Varanasi, India"}]}],"member":"320","published-online":{"date-parts":[[2024,11,21]]},"reference":[{"key":"e_1_3_1_2_2","volume-title":"Introduction to Public Health","author":"Schneider M.","year":"2011","unstructured":"M. Schneider. 2011. Introduction to Public Health. Sudbury, MA: Jones and Bartlett."},{"key":"e_1_3_1_3_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2017.01.034"},{"key":"e_1_3_1_4_2","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2007.1042"},{"key":"e_1_3_1_5_2","doi-asserted-by":"publisher","DOI":"10.1007\/s12652-021-03248-z"},{"key":"e_1_3_1_6_2","doi-asserted-by":"publisher","DOI":"10.1080\/00207721.2013.784372"},{"key":"e_1_3_1_7_2","doi-asserted-by":"publisher","DOI":"10.1109\/JSEN.2019.2946095"},{"key":"e_1_3_1_8_2","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2011.2129370"},{"key":"e_1_3_1_9_2","doi-asserted-by":"publisher","DOI":"10.1109\/TITB.2010.2087385"},{"key":"e_1_3_1_10_2","doi-asserted-by":"publisher","DOI":"10.1155\/2017\/9474806"},{"key":"e_1_3_1_11_2","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2011.2129370"},{"key":"e_1_3_1_12_2","first-page":"529","article-title":"Video based fall detection using features of motion, shape and histogram","author":"Albawendi Suad","year":"2018","unstructured":"Suad Albawendi, Ahmad Lot, Heather Powell, and Kofi Appiah. 2018. Video based fall detection using features of motion, shape and histogram. Proceeding of PETRA\u201918, (Corfu, Greece), 529\u2013536","journal-title":"Proceeding of PETRA\u201918, (Corfu, Greece)"},{"key":"e_1_3_1_13_2","doi-asserted-by":"publisher","DOI":"10.3390\/sym12050744"},{"key":"e_1_3_1_14_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.physa.2019.123205"},{"key":"e_1_3_1_15_2","first-page":"23\u201325 50\u201354","article-title":"Human fall detection algorithm based on YOLOv3","author":"Wang X.","year":"2020","unstructured":"X. Wang and K. Jia. 2020. Human fall detection algorithm based on YOLOv3. In Proceedings of the 2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC); (Qingdao, China). 23\u201325 50\u201354.","journal-title":"Proceedings of the 2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)"},{"key":"e_1_3_1_16_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.procs.2020.06.049"},{"key":"e_1_3_1_17_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3021795"},{"key":"e_1_3_1_18_2","first-page":"1350","article-title":"Multiple cameras fall dataset","author":"Auvinet Edouard","year":"2010","unstructured":"Edouard Auvinet, Caroline Rougier, Jean Meunier, Alain St-Arnaud, and Jacqueline Rousseau. 2010. Multiple cameras fall dataset. DIRO-Universit\u00b4e de Montr\u00b4eal, Tech. Rep, 1350.","journal-title":"DIRO-Universit\u00b4e de Montr\u00b4eal, Tech. Rep"},{"key":"e_1_3_1_19_2","doi-asserted-by":"publisher","DOI":"10.1109\/iEECON48109.2020.229491"},{"key":"e_1_3_1_20_2","doi-asserted-by":"publisher","DOI":"10.3390\/s18061918"},{"key":"e_1_3_1_21_2","doi-asserted-by":"publisher","DOI":"10.1080\/14639230410001684387"},{"key":"e_1_3_1_22_2","doi-asserted-by":"publisher","DOI":"10.1007\/s11760-013-0493-7"},{"key":"e_1_3_1_23_2","first-page":"1","article-title":"XCovNet: An optimized Xception convolutional neural network for classification of COVID-19 from point-of-care lung ultrasound images","author":"Madhu G.","year":"2023","unstructured":"G. Madhu, S. Kautish, Y. Gupta, G. Nagachandrika, S. M. Biju, and M. Kumar. 2023. XCovNet: An optimized Xception convolutional neural network for classification of COVID-19 from point-of-care lung ultrasound images. Multimedia Tools and Applications, 1\u201322.","journal-title":"Multimedia Tools and Applications"},{"key":"e_1_3_1_24_2","article-title":"Automated human tracing using gait and face using artificial neural network in surveillance system","author":"Kumar A.","year":"2022","unstructured":"A. Kumar, S. Jain, and M. Kumar. 2022. Automated human tracing using gait and face using artificial neural network in surveillance system. In International Conference on Data Analytics and Computing (277\u2013292). Singapore: Springer Nature Singapore.","journal-title":"International Conference on Data Analytics and Computing (277\u2013292)"},{"key":"e_1_3_1_25_2","first-page":"1","volume-title":"Multimedia Tools and Applications","author":"Nijhawan R.","year":"2023","unstructured":"R. Nijhawan, G. Sinha, A. Batra, M. Kumar, and H. Sharma. 2023. VTnet+ handcrafted based approach for food cuisines classification. Multimedia Tools and Applications, 1\u201321."},{"key":"e_1_3_1_26_2","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-018-6775-x"},{"key":"e_1_3_1_27_2","doi-asserted-by":"publisher","DOI":"10.1109\/CCECE58730.2023.10288941"},{"key":"e_1_3_1_28_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2022.118681"},{"key":"e_1_3_1_29_2","unstructured":"Image source: https:\/\/cocodataset.org\/"}],"container-title":["ACM Transactions on Asian and Low-Resource Language Information Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3687125","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3687125","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:10:21Z","timestamp":1750295421000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3687125"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,21]]},"references-count":28,"journal-issue":{"issue":"11","published-print":{"date-parts":[[2024,11,30]]}},"alternative-id":["10.1145\/3687125"],"URL":"https:\/\/doi.org\/10.1145\/3687125","relation":{},"ISSN":["2375-4699","2375-4702"],"issn-type":[{"value":"2375-4699","type":"print"},{"value":"2375-4702","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,21]]},"assertion":[{"value":"2023-09-10","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-08-01","order":2,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-11-21","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}