{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,15]],"date-time":"2026-03-15T23:05:17Z","timestamp":1773615917280,"version":"3.50.1"},"reference-count":36,"publisher":"Allerton Press","issue":"2","license":[{"start":{"date-parts":[[2025,4,1]],"date-time":"2025-04-01T00:00:00Z","timestamp":1743465600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,4,1]],"date-time":"2025-04-01T00:00:00Z","timestamp":1743465600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Aut. Control Comp. Sci."],"published-print":{"date-parts":[[2025,4]]},"DOI":"10.3103\/s014641162570021x","type":"journal-article","created":{"date-parts":[[2025,7,4]],"date-time":"2025-07-04T15:27:57Z","timestamp":1751642877000},"page":"255-265","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["DIAT-DSCNN-GRU-HARNet: A Lightweight DCNN for Video Based Classification of Human Activities"],"prefix":"10.3103","volume":"59","author":[{"family":"Ajay Waghumbare","sequence":"first","affiliation":[]},{"family":"Upasna Singh","sequence":"additional","affiliation":[]}],"member":"1627","published-online":{"date-parts":[[2025,7,4]]},"reference":[{"key":"7832_CR1","doi-asserted-by":"publisher","first-page":"214","DOI":"10.1016\/j.engappai.2018.08.009","volume":"76","author":"Zh. Tan","year":"2018","unstructured":"Tan, Zh., Xu, L., Zhong, W., Guo, X., and Wang, G., Online activity recognition and daily habit modeling for solitary elderly through indoor position-based stigmergy, Eng. Appl. Artif. Intell., 2018, vol. 76, pp. 214\u2013225. https:\/\/doi.org\/10.1016\/j.engappai.2018.08.009","journal-title":"Eng. Appl. Artif. Intell."},{"key":"7832_CR2","doi-asserted-by":"publisher","first-page":"14885","DOI":"10.1007\/s11042-020-08806-9","volume":"83","author":"M.A. Khan","year":"2020","unstructured":"Khan, M.A., Javed, K., Khan, S.A., Saba, T., Habib, U., Khan, J.A., and Abbasi, A.A., Human action recognition using fusion of multiview and deep features: An application to video surveillance, Multimedia Tools Appl., 2020, vol. 83, no. 5, pp. 14885\u201314911. https:\/\/doi.org\/10.1007\/s11042-020-08806-9","journal-title":"Multimedia Tools Appl."},{"key":"7832_CR3","doi-asserted-by":"publisher","unstructured":"Song, Z., Yin, Z., Yuan, Z., Zhang, Ch., Chi, W., Ling, Yo., and Zhang, Sh., Attention-oriented action recognition for real-time human\u2013robot interaction, 2020 25th International Conference on Pattern Recognition (ICPR), Milan, 2021, IEEE, 2021, pp. 7087\u20137094. https:\/\/doi.org\/10.1109\/icpr48806.2021.9412346","DOI":"10.1109\/icpr48806.2021.9412346"},{"key":"7832_CR4","doi-asserted-by":"publisher","unstructured":"Couturier, R., Noura, H.N., Salman, O., and Sider, A., A deep learning object detection method for an efficient clusters initialization, arXiv Preprint, 2021. https:\/\/doi.org\/10.48550\/arXiv.2104.13634","DOI":"10.48550\/arXiv.2104.13634"},{"key":"7832_CR5","doi-asserted-by":"publisher","first-page":"1272","DOI":"10.1109\/tcss.2021.3063318","volume":"8","author":"Sh. Zhong","year":"2021","unstructured":"Zhong, Sh., Bao, Z., Gong, Sh., and Xia, K., Person reidentification based on pose-invariant feature and B-KNN reranking, IEEE Trans. Comput. Soc. Syst., 2021, vol. 8, no. 5, pp. 1272\u20131281. https:\/\/doi.org\/10.1109\/tcss.2021.3063318","journal-title":"IEEE Trans. Comput. Soc. Syst."},{"key":"7832_CR6","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1109\/mnet.2019.1800310","volume":"33","author":"Ch. Dai","year":"2019","unstructured":"Dai, Ch., Liu, X., Lai, J., Li, P., and Chao, H.-C., Human behavior deep recognition architecture for smart city applications in the 5G environment, IEEE Network, 2019, vol. 33, no. 5, pp. 206\u2013211. https:\/\/doi.org\/10.1109\/mnet.2019.1800310","journal-title":"IEEE Network"},{"key":"7832_CR7","doi-asserted-by":"publisher","first-page":"161","DOI":"10.1016\/j.patrec.2018.07.011","volume":"112","author":"H. Kwon","year":"2018","unstructured":"Kwon, H., Kim, Ye., Lee, J.S., and Cho, M., First person action recognition via two-stream ConvNet with long-term fusion pooling, Pattern Recognit. Lett., 2018, vol. 112, pp. 161\u2013167. https:\/\/doi.org\/10.1016\/j.patrec.2018.07.011","journal-title":"Pattern Recognit. Lett."},{"key":"7832_CR8","doi-asserted-by":"publisher","first-page":"73971","DOI":"10.1109\/access.2023.3296252","volume":"11","author":"T. Shanableh","year":"2023","unstructured":"Shanableh, T., ViCo-MoCo-DL: Video coding and motion compensation solutions for human activity recognition using deep learning, IEEE Access, 2023, vol. 11, pp. 73971\u201373981. https:\/\/doi.org\/10.1109\/access.2023.3296252","journal-title":"IEEE Access"},{"key":"7832_CR9","doi-asserted-by":"publisher","first-page":"26901","DOI":"10.1007\/s11042-018-5893-9","volume":"77","author":"B. Meng","year":"2018","unstructured":"Meng, B., Liu, X., and Wang, X., Human action recognition based on quaternion spatial-temporal convolutional neural network and LSTM in RGB videos, Multimedia Tools Appl., 2018, vol. 77, no. 20, pp. 26901\u201326918. https:\/\/doi.org\/10.1007\/s11042-018-5893-9","journal-title":"Multimedia Tools Appl."},{"key":"7832_CR10","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1109\/tpami.2012.59","volume":"35","author":"Sh. Ji","year":"2013","unstructured":"Ji, Sh., Xu, W., Yang, M., and Yu, K., 3D convolutional neural networks for human action recognition, IEEE Trans. Pattern Anal. Mach. Intell., 2013, vol. 35, no. 1, pp. 221\u2013231. https:\/\/doi.org\/10.1109\/tpami.2012.59","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"7832_CR11","doi-asserted-by":"publisher","unstructured":"Donahue, J., Hendricks, L.A., Guadarrama, S., Rohrbach, M., Venugopalan, S., Saenko, K., and Darrell, T., Long-term recurrent convolutional networks for visual recognition and description, Technical Report UCB\/EECS-2014-180, Defense Technical Information Center, 2014. https:\/\/doi.org\/10.21236\/ada623249","DOI":"10.21236\/ada623249"},{"key":"7832_CR12","doi-asserted-by":"publisher","unstructured":"Wang, H. and Schmid, C., Action recognition with improved trajectories, 2013 IEEE International Conference on Computer Vision, Sydney, 2013, IEEE, 2013, pp. 3551\u20133558. https:\/\/doi.org\/10.1109\/iccv.2013.441","DOI":"10.1109\/iccv.2013.441"},{"key":"7832_CR13","doi-asserted-by":"publisher","unstructured":"Peng, X., Zou, Ch., Qiao, Yu., and Peng, Q., Action recognition with stacked Fisher vectors, Computer Vision\u2013ECCV 2014, Lecture Notes in Computer Science, vol. 8693, Cham: Springer, 2014, pp. 581\u2013595. https:\/\/doi.org\/10.1007\/978-3-319-10602-1_38","DOI":"10.1007\/978-3-319-10602-1_38"},{"key":"7832_CR14","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A. Krizhevsky","year":"2012","unstructured":"Krizhevsky, A., Sutskever, I., and Hinton, G.E., ImageNet classification with deep convolutional neural networks, Commun. ACM, 2012, vol. 60, no. 6, pp. 84\u201390. https:\/\/doi.org\/10.1145\/3065386","journal-title":"Commun. ACM"},{"key":"7832_CR15","doi-asserted-by":"publisher","unstructured":"Feichtenhofer, Ch., Pinz, A., and Wildes, R.P., Spatiotemporal multiplier networks for video action recognition, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, 2017, IEEE, 2017, pp. 7445\u20137454. https:\/\/doi.org\/10.1109\/cvpr.2017.787","DOI":"10.1109\/cvpr.2017.787"},{"key":"7832_CR16","doi-asserted-by":"publisher","unstructured":"Feichtenhofer, Ch., Pinz, A., and Wildes, R.P., Spatiotemporal multiplier networks for video action recognition, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, 2017, IEEE, 2016, pp. 3476\u20133484. https:\/\/doi.org\/10.1109\/cvpr.2017.787","DOI":"10.1109\/cvpr.2017.787"},{"key":"7832_CR17","unstructured":"Simonyan, K. and Zisserman, A., Two-stream convolutional networks for action recognition in videos, Proceedings of the 28th Conference on Neural Information Processing Systems (NIPS), Montreal, 2014, Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N.D., and Weinberger, K.Q., Eds., Cambridge, MA: MIT Press, 2014, vol. 1."},{"key":"7832_CR18","doi-asserted-by":"publisher","unstructured":"Wang, L., Xiong, Yu., Wang, Zh., Qiao, Yu., Lin, D., Tang, X., and Van Gool, L., Temporal segment networks: Towards good practices for deep action recognition, Computer Vision\u2013ECCV 2016, Leibe, B., Matas, J., Sebe, N., and Welling, M., Eds., Lecture Notes in Computer Science, vol. 9912, Cham: Springer, 2016, pp. 20\u201336. https:\/\/doi.org\/10.1007\/978-3-319-46484-8_2","DOI":"10.1007\/978-3-319-46484-8_2"},{"key":"7832_CR19","doi-asserted-by":"publisher","first-page":"29","DOI":"10.1145\/3397179","volume":"16","author":"Y. Li","year":"2020","unstructured":"Li, Y., Hu, H., Zhu, Z., and Zhou, G., SCANet: Sensor-based continuous authentication with two-stream convolutional neural networks, ACM Trans. Sen. Netw., 2020, vol. 16, no. 3, p. 29. https:\/\/doi.org\/10.1145\/3397179","journal-title":"ACM Trans. Sen. Netw."},{"key":"7832_CR20","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1109\/tpami.2012.59","volume":"35","author":"Sh. Ji","year":"2013","unstructured":"Ji, Sh., Xu, W., Yang, M., and Yu, K., 3D convolutional neural networks for human action recognition, IEEE Trans. Pattern Anal. Mach. Intell., 2013, vol. 35, no. 1, pp. 221\u2013231. https:\/\/doi.org\/10.1109\/tpami.2012.59","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"7832_CR21","doi-asserted-by":"publisher","unstructured":"Tran, D., Bourdev, L., Fergus, R., Torresani, L., and Paluri, M., Learning spatiotemporal features with 3D convolutional networks, 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 2015, IEEE, 2015, pp. 4489\u20134497. https:\/\/doi.org\/10.1109\/iccv.2015.510","DOI":"10.1109\/iccv.2015.510"},{"key":"7832_CR22","doi-asserted-by":"publisher","unstructured":"Ballas, N., Yao, L., Pal, C., and Courville, A., Delving deeper into convolutional networks for learning video representations, arXiv Preprint, 2015. https:\/\/doi.org\/10.48550\/arXiv.1511.06432","DOI":"10.48550\/arXiv.1511.06432"},{"key":"7832_CR23","unstructured":"Mnih, V., Heess, N., Graves, A., and Kavukcuoglu, K., Recurrent models of visual attention, Advances in Neural Information Processing Systems, Montreal, 2014, Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N., and Weinberger, K.Q., Eds., Curran Associates, 2014, vol. 27. https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2014\/file\/3e456b31302cf8210edd4029292a40ad-Paper.pdf."},{"key":"7832_CR24","doi-asserted-by":"publisher","unstructured":"Chen, Yu., Zhong, K., Zhang, J., Sun, Q., and Zhao, X., LSTM networks for mobile human activity recognition, Proceedings of the 2016 International Conference on Artificial Intelligence: Technologies and Applications, Advances in Intelligent Systems Research, Atlantis Press, 2016. https:\/\/doi.org\/10.2991\/icaita-16.2016.13","DOI":"10.2991\/icaita-16.2016.13"},{"key":"7832_CR25","doi-asserted-by":"publisher","first-page":"2816","DOI":"10.3390\/s23052816","volume":"23","author":"R. Vrskova","year":"2023","unstructured":"Vrskova, R., Kamencay, P., Hudec, R., and Sykora, P., A new deep-learning method for human activity recognition, Sensors, 2023, vol. 23, no. 5, p. 2816. https:\/\/doi.org\/10.3390\/s23052816","journal-title":"Sensors"},{"key":"7832_CR26","doi-asserted-by":"publisher","first-page":"702","DOI":"10.1134\/s1054661819040084","volume":"29","author":"A. Khelalef","year":"2019","unstructured":"Khelalef, A., Ababsa, F., and Benoudjit, N., An efficient human activity recognition technique based on deep learning, Patern Recognit. Image Anal., 2019, vol. 29, no. 4, pp. 702\u2013715. https:\/\/doi.org\/10.1134\/s1054661819040084","journal-title":"Patern Recognit. Image Anal."},{"key":"7832_CR27","doi-asserted-by":"publisher","unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., and Adam, H., MobileNets: Efficient convolutional neural networks for mobile vision applications, arXiv Preprint, 2017. https:\/\/doi.org\/10.48550\/arXiv.1704.04861","DOI":"10.48550\/arXiv.1704.04861"},{"key":"7832_CR28","doi-asserted-by":"publisher","unstructured":"Szegedy, Ch., Vanhoucke, V., Ioffe, S., Shlens, J., and Wojna, Z., Rethinking the inception architecture for computer vision, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, 2016, IEEE, 2016, pp. 2818\u20132826. https:\/\/doi.org\/10.1109\/cvpr.2016.308","DOI":"10.1109\/cvpr.2016.308"},{"key":"7832_CR29","doi-asserted-by":"publisher","unstructured":"Chollet, F., Xception: Deep learning with depthwise separable convolutions, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, 2017, IEEE, 2017, pp. 1800\u20131807. https:\/\/doi.org\/10.1109\/cvpr.2017.195","DOI":"10.1109\/cvpr.2017.195"},{"key":"7832_CR30","doi-asserted-by":"publisher","unstructured":"Yu, F. and Koltun, V., Multi-scale context aggregation by dilated convolutions, arXiv Preprint, 2015. https:\/\/doi.org\/10.48550\/arXiv.1511.07122","DOI":"10.48550\/arXiv.1511.07122"},{"key":"7832_CR31","doi-asserted-by":"publisher","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J., Deep residual learning for image recognition, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, 2016, IEEE, 2016, pp. 770\u2013778. https:\/\/doi.org\/10.1109\/CVPR.2016.90","DOI":"10.1109\/CVPR.2016.90"},{"key":"7832_CR32","doi-asserted-by":"publisher","unstructured":"Martinez, J., Black, M.J., and Romero, J., On human motion prediction using recurrent neural networks, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, 2017, IEEE, 2017, pp.\u00a04674\u20134683. https:\/\/doi.org\/10.1109\/cvpr.2017.497","DOI":"10.1109\/cvpr.2017.497"},{"key":"7832_CR33","doi-asserted-by":"publisher","unstructured":"Oreifej, O. and Liu, Z., HON4D: Histogram of oriented 4D normals for activity recognition from depth sequences, 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, 2013, IEEE, 2013, pp. 716\u2013723. https:\/\/doi.org\/10.1109\/cvpr.2013.98","DOI":"10.1109\/cvpr.2013.98"},{"key":"7832_CR34","unstructured":"CRCV, Center for Research in Computer Vision at the University of Central Florida, 2023. https:\/\/www.crcv.ucf.edu\/data\/UCF-ARG.php. Cited December 15, 2023."},{"key":"7832_CR35","doi-asserted-by":"publisher","unstructured":"Huang, G., Liu, Zh., Van Der Maaten, L., and Weinberger, K.Q., Densely connected convolutional networks, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, 2017, IEEE, 2017, pp.\u00a02261\u20132269. https:\/\/doi.org\/10.1109\/cvpr.2017.243","DOI":"10.1109\/cvpr.2017.243"},{"key":"7832_CR36","doi-asserted-by":"publisher","unstructured":"Simonyan, K. and Zisserman, A., Very deep convolutional networks for large-scale image recognition, arXiv Preprint, 2014. https:\/\/doi.org\/10.48550\/arXiv.1409.1556","DOI":"10.48550\/arXiv.1409.1556"}],"container-title":["Automatic Control and Computer Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.3103\/S014641162570021X.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.3103\/S014641162570021X","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.3103\/S014641162570021X.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,15]],"date-time":"2026-03-15T22:07:07Z","timestamp":1773612427000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.3103\/S014641162570021X"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4]]},"references-count":36,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2025,4]]}},"alternative-id":["7832"],"URL":"https:\/\/doi.org\/10.3103\/s014641162570021x","relation":{},"ISSN":["0146-4116","1558-108X"],"issn-type":[{"value":"0146-4116","type":"print"},{"value":"1558-108X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,4]]},"assertion":[{"value":"19 January 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 June 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 June 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 July 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"The authors of this work declare that they have no conflicts of interest.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"CONFLICT OF INTEREST"}}]}}