{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T14:31:30Z","timestamp":1775745090465,"version":"3.50.1"},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2024,1,10]],"date-time":"2024-01-10T00:00:00Z","timestamp":1704844800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,10]],"date-time":"2024-01-10T00:00:00Z","timestamp":1704844800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Comput &amp; Applic"],"published-print":{"date-parts":[[2024,4]]},"DOI":"10.1007\/s00521-023-09362-7","type":"journal-article","created":{"date-parts":[[2024,1,10]],"date-time":"2024-01-10T15:02:45Z","timestamp":1704898965000},"page":"5515-5528","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":18,"title":["Modeling transformer architecture with attention layer for human activity recognition"],"prefix":"10.1007","volume":"36","author":[{"given":"Gunjan","family":"Pareek","sequence":"first","affiliation":[]},{"given":"Swati","family":"Nigam","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4022-9945","authenticated-orcid":false,"given":"Rajiv","family":"Singh","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,1,10]]},"reference":[{"key":"9362_CR1","doi-asserted-by":"publisher","first-page":"13547","DOI":"10.1007\/s00521-023-08363-w","volume":"35","author":"L D\u2019Arco","year":"2023","unstructured":"D\u2019Arco L, Wang H, Zheng H (2023) DeepHAR: a deep feed-forward neural network algorithm for smart insole-based human activity recognition. Neural Comput Appl 35:13547\u201313563. https:\/\/doi.org\/10.1007\/s00521-023-08363-w","journal-title":"Neural Comput Appl"},{"key":"9362_CR2","doi-asserted-by":"publisher","first-page":"13321","DOI":"10.1007\/s00521-023-08440-0","volume":"35","author":"A Kushwaha","year":"2023","unstructured":"Kushwaha A, Khare A, Prakash O (2023) Micro-network-based deep convolutional neural network for human activity recognition from realistic and multi-view visual data. Neural Comput Appl 35:13321\u201313341. https:\/\/doi.org\/10.1007\/s00521-023-08440-0","journal-title":"Neural Comput Appl"},{"key":"9362_CR3","doi-asserted-by":"publisher","first-page":"14624","DOI":"10.1038\/s41598-023-39744-9","volume":"13","author":"HP Nguyen","year":"2023","unstructured":"Nguyen HP, Ribeiro B (2023) Video action recognition collaborative learning with dynamics via PSO-ConvNet Transformer. Sci Rep 13:14624. https:\/\/doi.org\/10.1038\/s41598-023-39744-9","journal-title":"Sci Rep"},{"key":"9362_CR4","doi-asserted-by":"publisher","DOI":"10.1016\/j.sciaf.2023.e01796","volume":"21","author":"EM Saoudi","year":"2023","unstructured":"Saoudi EM, Jaafari J, Andaloussi SJ (2023) Advancing human action recognition: a hybrid approach using attention-based LSTM and 3D CNN. Sci Afr 21:e01796. https:\/\/doi.org\/10.1016\/j.sciaf.2023.e01796","journal-title":"Sci Afr"},{"key":"9362_CR5","doi-asserted-by":"publisher","DOI":"10.3390\/s23146384","author":"GA Surek","year":"2023","unstructured":"Surek GA, Seman LO, Stefenon SF, Mariani VC, Coelho LD (2023) Video-based human activity recognition using deep learning approaches. Sensors. https:\/\/doi.org\/10.3390\/s23146384","journal-title":"Sensors"},{"key":"9362_CR6","doi-asserted-by":"publisher","first-page":"222","DOI":"10.1049\/cvi2.12154","volume":"17","author":"H Zhang","year":"2023","unstructured":"Zhang H, Wang L, Sun J (2023) Exploiting spatio-temporal knowledge for video action recognition. IET Comput Vision 17:222\u2013230. https:\/\/doi.org\/10.1049\/cvi2.12154","journal-title":"IET Comput Vision"},{"key":"9362_CR7","doi-asserted-by":"publisher","first-page":"8273546","DOI":"10.1155\/2023\/8273546","volume":"2023","author":"S Zhu","year":"2023","unstructured":"Zhu S, Chen W, Liu F, Zhang X, Han X (2023) Human activity recognition based on a modified capsule network. Mob Inf Syst 2023:8273546. https:\/\/doi.org\/10.1155\/2023\/8273546","journal-title":"Mob Inf Syst"},{"issue":"7","key":"9362_CR8","doi-asserted-by":"publisher","first-page":"5427","DOI":"10.1007\/s11831-022-09772-1","volume":"29","author":"B Tyagi","year":"2022","unstructured":"Tyagi B, Nigam S, Singh R (2022) A review of deep learning techniques for crowd behavior analysis. Arch Comput Method Eng 29(7):5427\u20135455","journal-title":"Arch Comput Method Eng"},{"key":"9362_CR9","doi-asserted-by":"publisher","unstructured":"Umar IM, Ibrahim KM, Gital AYU, Zambuk FU, Lawal MA, Yakubu ZI (2022) Hybrid model for human activity recognition using an inflated i3-D two stream convolutional-LSTM network with optical flow mechanism. In: 2022 IEEE Delhi section conference, DELCON 2022. https:\/\/doi.org\/10.1109\/DELCON54057.2022.9752782.","DOI":"10.1109\/DELCON54057.2022.9752782"},{"issue":"5","key":"9362_CR10","doi-asserted-by":"publisher","first-page":"4707","DOI":"10.1007\/s12652-022-04374-y","volume":"14","author":"S Nigam","year":"2023","unstructured":"Nigam S, Singh R, Singh MK, Singh VK (2023) Multiview human activity recognition using uniform rotation invariant local binary patterns. J Ambient Intell Humaniz Comput 14(5):4707\u20134725","journal-title":"J Ambient Intell Humaniz Comput"},{"key":"9362_CR11","doi-asserted-by":"publisher","unstructured":"Manaf FA, Singh S (2021) A novel hybridization model for human activity recognition using stacked parallel LSTMs with 2D-CNN for feature extraction. In: 2021 12th International conference on computing communication and networking technologies (ICCCNT), pp 1\u20137. https:\/\/doi.org\/10.1109\/ICCCNT51525.2021.9579686","DOI":"10.1109\/ICCCNT51525.2021.9579686"},{"key":"9362_CR12","first-page":"831","volume":"26","author":"S Nigam","year":"2019","unstructured":"Nigam S, Singh R, Misra AK (2019) A review of computational approaches for human behavior detection. Arch Comput Method Eng 26:831\u2013863","journal-title":"Arch Comput Method Eng"},{"key":"9362_CR13","doi-asserted-by":"publisher","first-page":"1","DOI":"10.3390\/s19143160","volume":"19","author":"I Rodr\u00edguez-Moreno","year":"2019","unstructured":"Rodr\u00edguez-Moreno I, Mart\u00ednez-Otzeta JM, Sierra B, Rodriguez I, Jauregi E (2019) Video activity recognition: state-of-the-art. Sensors (Switzerland) 19:1\u201325. https:\/\/doi.org\/10.3390\/s19143160","journal-title":"Sensors (Switzerland)"},{"key":"9362_CR14","doi-asserted-by":"publisher","first-page":"56855","DOI":"10.1109\/ACCESS.2020.2982225","volume":"8","author":"K Xia","year":"2020","unstructured":"Xia K, Huang J, Wang H (2020) LSTM-CNN architecture for human activity recognition. IEEE Access 8:56855\u201356866. https:\/\/doi.org\/10.1109\/ACCESS.2020.2982225","journal-title":"IEEE Access"},{"key":"9362_CR15","doi-asserted-by":"publisher","unstructured":"Fereidoonian F, Firouzi F, Farahani B (2020) Human Activity recognition: from sensors to applications. In: 2020 International conference on omni-layer intelligent systems, COINS 2020. https:\/\/doi.org\/10.1109\/COINS49042.2020.9191417","DOI":"10.1109\/COINS49042.2020.9191417"},{"key":"9362_CR16","doi-asserted-by":"publisher","first-page":"60736","DOI":"10.1109\/ACCESS.2019.2913393","volume":"7","author":"M Ehatisham-Ul-Haq","year":"2019","unstructured":"Ehatisham-Ul-Haq M, Javed A, Azam MA, Malik HMA, Irtaza A, Lee IH, Mahmood MT (2019) Robust human activity recognition using multimodal feature-level fusion. IEEE Access 7:60736\u201360751. https:\/\/doi.org\/10.1109\/ACCESS.2019.2913393","journal-title":"IEEE Access"},{"key":"9362_CR17","doi-asserted-by":"publisher","first-page":"164453","DOI":"10.1109\/ACCESS.2020.3022287","volume":"8","author":"M Muaaz","year":"2020","unstructured":"Muaaz M, Chelli A, Abdelgawwad AA, Mallofr\u00e9 AC, P\u00e4tzold M (2020) WiWeHAR: Multimodal human activity recognition using Wi-Fi and wearable sensing modalities. IEEE Access 8:164453\u2013164470. https:\/\/doi.org\/10.1109\/ACCESS.2020.3022287","journal-title":"IEEE Access"},{"key":"9362_CR18","doi-asserted-by":"publisher","unstructured":"Karpathy A, Toderici G, Shetty S, Leung T, Sukthankar R, Li FF (2014) Large-scale video classification with convolutional neural networks. In: Proceedings of the IEEE computer society conference on computer vision and pattern recognition, pp 1725\u20131732. https:\/\/doi.org\/10.1109\/CVPR.2014.223","DOI":"10.1109\/CVPR.2014.223"},{"key":"9362_CR19","unstructured":"Soomro K, Zamir AR, Shah M (2012) UCF101: A Dataset of 101 human actions classes from videos in the wild"},{"key":"9362_CR20","doi-asserted-by":"publisher","first-page":"634","DOI":"10.1109\/TMM.2017.2749159","volume":"20","author":"X Wang","year":"2018","unstructured":"Wang X, Gao L, Wang P, Sun X, Liu X (2018) Two-stream 3-D convNet fusion for action recognition in videos with arbitrary size and length. IEEE Trans Multimedia 20:634\u2013644. https:\/\/doi.org\/10.1109\/TMM.2017.2749159","journal-title":"IEEE Trans Multimedia"},{"key":"9362_CR21","unstructured":"Wang L, Xiong Y, Wang Z, Qiao Y (2015) Towards good practices for very deep two-stream ConvNets, pp 1\u20135"},{"key":"9362_CR22","doi-asserted-by":"publisher","unstructured":"Wang L, Qiao Y, Tang X (2015) Action recognition with trajectory-pooled deep-convolutional descriptors. In: Proceedings of the IEEE computer society conference on computer vision and pattern recognition, pp 4305\u20134314. https:\/\/doi.org\/10.1109\/CVPR.2015.7299059","DOI":"10.1109\/CVPR.2015.7299059"},{"key":"9362_CR23","doi-asserted-by":"publisher","unstructured":"Feichtenhofer C, Pinz A, Zisserman A (2016) Convolutional two-stream network fusion for video action recognition. In: Proceedings of the IEEE computer society conference on computer vision and pattern recognition. 2016-Decem, pp 1933\u20131941. https:\/\/doi.org\/10.1109\/CVPR.2016.213","DOI":"10.1109\/CVPR.2016.213"},{"key":"9362_CR24","doi-asserted-by":"publisher","unstructured":"Bilen H, Fernando B, Gavves E, Vedaldi A, Gould S(2016) Dynamic image networks for action recognition. In: Proceedings of the IEEE computer society conference on computer vision and pattern recognition. 2016-Decem, pp 3034\u20133042. https:\/\/doi.org\/10.1109\/CVPR.2016.331","DOI":"10.1109\/CVPR.2016.331"},{"key":"9362_CR25","doi-asserted-by":"publisher","unstructured":"Carreira J, Zisserman A (2017) Quo Vadis, action recognition? A new model and the kinetics dataset. In: Proceedings\u201430th IEEE conference on computer vision and pattern recognition, CVPR 2017. 2017-Janua, pp 4724\u20134733. https:\/\/doi.org\/10.1109\/CVPR.2017.502","DOI":"10.1109\/CVPR.2017.502"},{"key":"9362_CR26","doi-asserted-by":"publisher","first-page":"1155","DOI":"10.1109\/ACCESS.2017.2778011","volume":"6","author":"A Ullah","year":"2017","unstructured":"Ullah A, Ahmad J, Muhammad K, Sajjad M, Baik SW (2017) Action recognition in video sequences using deep bi-directional LSTM with CNN features. IEEE Access 6:1155\u20131166. https:\/\/doi.org\/10.1109\/ACCESS.2017.2778011","journal-title":"IEEE Access"},{"key":"9362_CR27","doi-asserted-by":"publisher","unstructured":"Chen J, Xu Y, Zhang C, Xu Z, Meng X, Wang J (2019) An improved two-stream 3D convolutional neural network for human action recognition. In: 2019 25th International conference on automation and computing (ICAC), pp 1\u20136. https:\/\/doi.org\/10.23919\/IConAC.2019.8894962","DOI":"10.23919\/IConAC.2019.8894962"},{"key":"9362_CR28","doi-asserted-by":"publisher","first-page":"19799","DOI":"10.1109\/ACCESS.2020.2968529","volume":"8","author":"S Tanberk","year":"2020","unstructured":"Tanberk S, Kilimci ZH, Tukel DB, Uysal M, Akyokus S (2020) A hybrid deep model using deep learning and dense optical flow approaches for human activity recognition. IEEE Access 8:19799\u201319809. https:\/\/doi.org\/10.1109\/ACCESS.2020.2968529","journal-title":"IEEE Access"},{"key":"9362_CR29","doi-asserted-by":"publisher","unstructured":"Gatt T, Seychell D, Dingli A (2019) Detecting human abnormal behaviour through a video generated model. In: International symposium on image and signal processing and analysis, ISPA. 2019-Septe, pp 264\u2013270. https:\/\/doi.org\/10.1109\/ISPA.2019.8868795","DOI":"10.1109\/ISPA.2019.8868795"},{"key":"9362_CR30","doi-asserted-by":"publisher","unstructured":"Zheng Y, Liu Q, Chen E, Ge Y, Zhao JL (2014) Time series classification using multi-channels deep convolutional neural networks. In: Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). 8485 LNCS, pp 298\u2013310. https:\/\/doi.org\/10.1007\/978-3-319-08010-9_33","DOI":"10.1007\/978-3-319-08010-9_33"},{"key":"9362_CR31","doi-asserted-by":"publisher","DOI":"10.3390\/s16010115","author":"FJ Ord\u00f3\u00f1ez","year":"2016","unstructured":"Ord\u00f3\u00f1ez FJ, Roggen D (2016) Deep convolutional and LSTM recurrent neural networks for multimodal wearable activity recognition. Sensors. https:\/\/doi.org\/10.3390\/s16010115","journal-title":"Sensors"},{"key":"9362_CR32","doi-asserted-by":"publisher","unstructured":"Mishra P, Dey S, Ghosh SS, Seal DB, Goswami S (2019) Human Activity Recognition using Deep Neural Network. In: 2019 International conference on data science and engineering (ICDSE). pp. 77\u201383. https:\/\/doi.org\/10.1109\/ICDSE47409.2019.8971476","DOI":"10.1109\/ICDSE47409.2019.8971476"},{"key":"9362_CR33","doi-asserted-by":"publisher","unstructured":"Khimraj, Shukla, PK, Vijayvargiya A, Kumar R (2020) Human Activity Recognition using Accelerometer and Gyroscope Data from Smartphones. In: Proceedings - 2020 international conference on emerging trends in communication, control and computing, ICONC3 2020. https:\/\/doi.org\/10.1109\/ICONC345789.2020.9117456","DOI":"10.1109\/ICONC345789.2020.9117456"},{"key":"9362_CR34","doi-asserted-by":"publisher","unstructured":"Kuehne H, Jhuang H, Garrote E, Poggio T, Serre T (2011) HMDB: A large video database for human motion recognition. In: Proceedings of the IEEE international conference on computer vision, pp 2556\u20132563. https:\/\/doi.org\/10.1109\/ICCV.2011.6126543","DOI":"10.1109\/ICCV.2011.6126543"},{"key":"9362_CR35","doi-asserted-by":"publisher","unstructured":"Chen C, Jafari R, Kehtarnavaz, N (2016) Fusion of depth, skeleton, and inertial data for human action recognition. In: 2016 IEEE International conference on acoustics, speech and signal processing (ICASSP). pp. 2712\u20132716. https:\/\/doi.org\/10.1109\/ICASSP.2016.7472170","DOI":"10.1109\/ICASSP.2016.7472170"},{"key":"9362_CR36","doi-asserted-by":"publisher","unstructured":"Li K, Zhao X, Bian J, Tan M (2017) Sequential learning for multimodal 3D human activity recognition with Long-Short Term Memory. IN: 2017 IEEE International conference on mechatronics and automation, ICMA 2017, pp 1556\u20131561. https:\/\/doi.org\/10.1109\/ICMA.2017.8016048","DOI":"10.1109\/ICMA.2017.8016048"},{"key":"9362_CR37","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-93000-8_42","volume-title":"Human action recognition using fusion of depth and inertial sensors","author":"Z Fuad","year":"2018","unstructured":"Fuad Z, Unel M (2018) Human action recognition using fusion of depth and inertial sensors. Springer, Berlin. https:\/\/doi.org\/10.1007\/978-3-319-93000-8_42"},{"key":"9362_CR38","doi-asserted-by":"publisher","first-page":"9324","DOI":"10.1109\/JSEN.2018.2869807","volume":"18","author":"A Manzi","year":"2018","unstructured":"Manzi A, Moschetti A, Limosani R, Fiorini L, Cavallo F (2018) Enhancing activity recognition of self-localized robot through depth camera and wearable sensors. IEEE Sens J 18:9324\u20139331. https:\/\/doi.org\/10.1109\/JSEN.2018.2869807","journal-title":"IEEE Sens J"},{"key":"9362_CR39","doi-asserted-by":"publisher","unstructured":"Sefen B, Baumbach S, Dengel A, Abdennadher S (2016) Human activity recognition using sensor data of smartphones and smartwatches. In: ICAART 2016 - Proceedings of the 8th international conference on agents and artificial intelligence. 2, pp 488\u2013493. https:\/\/doi.org\/10.5220\/0005816004880493","DOI":"10.5220\/0005816004880493"},{"key":"9362_CR40","doi-asserted-by":"publisher","first-page":"857","DOI":"10.1109\/TMC.2018.2841905","volume":"18","author":"P Bharti","year":"2019","unstructured":"Bharti P, De D, Chellappan S, Das SK (2019) HuMAn: Complex activity recognition with multi-modal multi-positional body sensing. IEEE Trans Mob Comput 18:857\u2013870. https:\/\/doi.org\/10.1109\/TMC.2018.2841905","journal-title":"IEEE Trans Mob Comput"},{"key":"9362_CR41","doi-asserted-by":"publisher","unstructured":"Martiez-Gonzalez A, Villamizar M, Canevet O, Odobez JM (2018) Real-time convolutional networks for depth-based human pose estimation. In: IEEE International conference on intelligent robots and systems, pp 41\u201347. https:\/\/doi.org\/10.1109\/IROS.2018.8593383","DOI":"10.1109\/IROS.2018.8593383"},{"key":"9362_CR42","doi-asserted-by":"publisher","unstructured":"Mohammad AN, Ohashi H, Ahmed S, Nakamura K, Akiyama T, Sato T, Nguyen P, Dengel A (2018) Hierarchical model for zero-shot activity recognition using wearable sensors. In: ICAART 2018 - Proceedings of the 10th international conference on agents and artificial intelligence. 2, pp 478\u2013485. https:\/\/doi.org\/10.5220\/0006595204780485","DOI":"10.5220\/0006595204780485"},{"key":"9362_CR43","doi-asserted-by":"publisher","unstructured":"Cruciani F, Sun C, Zhang S, Nugent C, Li C, Song S, Cheng C, Cleland I, McCullagh P (2019) A public domain dataset for human activity recognition in free-living conditions. In: Proceedings - 2019 IEEE SmartWorld, ubiquitous intelligence and computing, advanced and trusted computing, scalable computing and communications, internet of people and smart city innovation, SmartWorld\/UIC\/ATC\/SCALCOM\/IOP\/SCI 2019, pp 166\u2013171. https:\/\/doi.org\/10.1109\/SmartWorld-UIC-ATC-SCALCOM-IOP-SCI.2019.00071","DOI":"10.1109\/SmartWorld-UIC-ATC-SCALCOM-IOP-SCI.2019.00071"},{"key":"9362_CR44","doi-asserted-by":"publisher","first-page":"2033","DOI":"10.1016\/j.patrec.2012.12.014","volume":"34","author":"R Chavarriaga","year":"2013","unstructured":"Chavarriaga R, Sagha H, Calatroni A, Digumarti ST, Tr\u00f6ster G, Mill\u00e1n JDR, Roggen D (2013) The opportunity challenge: a benchmark database for on-body sensor-based activity recognition. Pattern Recogn Lett 34:2033\u20132042. https:\/\/doi.org\/10.1016\/j.patrec.2012.12.014","journal-title":"Pattern Recogn Lett"},{"key":"9362_CR45","doi-asserted-by":"publisher","unstructured":"Chen C, Jafari R, Kehtarnavaz N (2015) UTD-MHAD: A multimodal dataset for human action recognition utilizing a depth camera and a wearable inertial sensor. In: Proceedings - international conference on image processing, ICIP. 2015-Decem, pp 168\u2013172. https:\/\/doi.org\/10.1109\/ICIP.2015.7350781","DOI":"10.1109\/ICIP.2015.7350781"},{"key":"9362_CR46","doi-asserted-by":"crossref","unstructured":"Nigam S, Singh R, Singh MK, Singh VK (2021) Multiple views-based recognition of human activities using uniform patterns. In: 2021 Sixth international conference on image information processing (ICIIP), Vol. 6, pp. 483\u2013488. IEEE","DOI":"10.1109\/ICIIP53038.2021.9702576"},{"key":"9362_CR47","doi-asserted-by":"publisher","unstructured":"Ofli F, Chaudhry R, Kurillo G, Vidal R, Bajcsy R (2013) Berkeley MHAD: A comprehensive Multimodal Human Action Database. In: Proceedings of IEEE Workshop on applications of computer vision, pp 53\u201360. https:\/\/doi.org\/10.1109\/WACV.2013.6474999","DOI":"10.1109\/WACV.2013.6474999"},{"key":"9362_CR48","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s42979-020-00169-0","volume":"1","author":"DG Shreyas","year":"2020","unstructured":"Shreyas DG, Raksha S, Prasad BG (2020) Implementation of an anomalous human activity recognition system. SN Comput Sci 1:1\u201310. https:\/\/doi.org\/10.1007\/s42979-020-00169-0","journal-title":"SN Comput Sci"},{"key":"9362_CR49","doi-asserted-by":"publisher","DOI":"10.1145\/3447744","author":"K Chen","year":"2021","unstructured":"Chen K, Zhang D, Yao L, Guo B, Yu Z, Liu Y (2021) Deep learning for sensor-based human activity recognition: overview, challenges, and opportunities. ACM Comput Surv. https:\/\/doi.org\/10.1145\/3447744","journal-title":"ACM Comput Surv"},{"key":"9362_CR50","doi-asserted-by":"publisher","DOI":"10.1155\/2018\/8580959","author":"J Sun","year":"2018","unstructured":"Sun J, Fu Y, Li S, He J, Xu C, Tan L (2018) Sequential human activity recognition based on deep convolutional network and extreme learning machine using wearable sensors. J Sens. https:\/\/doi.org\/10.1155\/2018\/8580959","journal-title":"J Sens"},{"key":"9362_CR51","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2021.106970","volume":"223","author":"SK Yadav","year":"2021","unstructured":"Yadav SK, Tiwari K, Pandey HM, Akbar SA (2021) A review of multimodal human activity recognition with special emphasis on classification, applications, challenges and future directions. Knowl-Based Syst 223:106970. https:\/\/doi.org\/10.1016\/j.knosys.2021.106970","journal-title":"Knowl-Based Syst"},{"key":"9362_CR52","doi-asserted-by":"publisher","unstructured":"Kalfaoglu ME, Kalkan S, Alatan AA (2020) Late Temporal Modeling in 3D CNN Architectures with BERT for Action Recognition. In: Lecture notes in computer science (including subseries lecture notes in artificial intelligence and lecture notes in bioinformatics). 12539 LNCS, pp 731\u2013747. https:\/\/doi.org\/10.1007\/978-3-030-68238-5_48","DOI":"10.1007\/978-3-030-68238-5_48"}],"container-title":["Neural Computing and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-023-09362-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00521-023-09362-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-023-09362-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,8]],"date-time":"2024-03-08T21:58:27Z","timestamp":1709935107000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00521-023-09362-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1,10]]},"references-count":52,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2024,4]]}},"alternative-id":["9362"],"URL":"https:\/\/doi.org\/10.1007\/s00521-023-09362-7","relation":{},"ISSN":["0941-0643","1433-3058"],"issn-type":[{"value":"0941-0643","type":"print"},{"value":"1433-3058","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,1,10]]},"assertion":[{"value":"8 April 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 December 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 January 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that there is no conflict of interest regarding this manuscript and received no funding for this work.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}