{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,16]],"date-time":"2026-02-16T10:00:55Z","timestamp":1771236055528,"version":"3.50.1"},"reference-count":41,"publisher":"MDPI AG","issue":"21","license":[{"start":{"date-parts":[[2022,11,2]],"date-time":"2022-11-02T00:00:00Z","timestamp":1667347200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Ministry of Agriculture, Food and Rural Affairs","award":["322063-03-1-SB010"],"award-info":[{"award-number":["322063-03-1-SB010"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"<jats:p>In order to provide intelligent and efficient healthcare services in the Internet of Medical Things (IoMT), human action recognition (HAR) can play a crucial role. As a result of their stringent requirements, such as high computational complexity and memory efficiency, classical HAR techniques are not applicable to modern and intelligent healthcare services, e.g., IoMT. To address these issues, we present in this paper a novel HAR technique for healthcare services in IoMT. This model, referred to as the spatio-temporal graph convolutional network (STGCN), primarily aims at skeleton-based human\u2013machine interfaces. By independently extracting spatial and temporal features, STGCN significantly reduces information loss. Spatio-temporal information is extracted independently of the exact spatial and temporal point, ensuring the extraction of useful features for HAR. Using only joint data and fewer parameters, we demonstrate that our proposed STGCN achieved 92.2% accuracy on the skeleton dataset. Unlike multi-channel methods, which use a combination of joint and bone data and have a large number of parameters, multi-channel methods use both joint and bone data. As a result, STGCN offers a good balance between accuracy, memory consumption, and processing time, making it suitable for detecting medical conditions.<\/jats:p>","DOI":"10.3390\/s22218438","type":"journal-article","created":{"date-parts":[[2022,11,3]],"date-time":"2022-11-03T03:53:07Z","timestamp":1667447587000},"page":"8438","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":8,"title":["A Spatio-Temporal Graph Convolutional Network Model for Internet of Medical Things (IoMT)"],"prefix":"10.3390","volume":"22","author":[{"given":"Dipon Kumar","family":"Ghosh","sequence":"first","affiliation":[{"name":"Department of Computer Science and Engineering, Brac University, Dhaka 1212, Bangladesh"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0306-4029","authenticated-orcid":false,"given":"Amitabha","family":"Chakrabarty","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering, Brac University, Dhaka 1212, Bangladesh"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7668-3838","authenticated-orcid":false,"given":"Hyeonjoon","family":"Moon","sequence":"additional","affiliation":[{"name":"Depatment of Computer Science and Engineering, Sejong University, Seoul 05006, Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3229-6785","authenticated-orcid":false,"given":"M. Jalil","family":"Piran","sequence":"additional","affiliation":[{"name":"Depatment of Computer Science and Engineering, Sejong University, Seoul 05006, Korea"}]}],"member":"1968","published-online":{"date-parts":[[2022,11,2]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"e4049","DOI":"10.1002\/ett.4049","article-title":"A survey on security threats and countermeasures in internet of medical things (IoMT)","volume":"33","author":"Papaioannou","year":"2022","journal-title":"Trans. Emerg. Telecommun. Technol."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"16921","DOI":"10.1109\/JIOT.2021.3053420","article-title":"Deep learning methods in internet of medical things for valvular heart disease screening system","volume":"8","author":"Su","year":"2021","journal-title":"IEEE Internet Things J."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"8864","DOI":"10.1109\/TII.2021.3128954","article-title":"An intelligent trust cloud management method for secure clustering in 5G enabled internet of medical things","volume":"18","author":"Yang","year":"2021","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"139","DOI":"10.1016\/j.comcom.2020.05.048","article-title":"An effective feature engineering for DNN using hybrid PCA-GWO for intrusion detection in IoMT architecture","volume":"160","author":"Maddikunta","year":"2020","journal-title":"Comput. Commun."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"1977","DOI":"10.1109\/JBHI.2021.3112693","article-title":"On the Design of Blockchain-Based ECDSA With Fault-Tolerant Batch Verification Protocol for Blockchain-Enabled IoMT","volume":"26","author":"Xiong","year":"2022","journal-title":"IEEE J. Biomed. Health Inform."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"1642","DOI":"10.1007\/s00415-017-8424-0","article-title":"Freezing of gait and fall detection in Parkinson\u2019s disease using wearable sensors: A systematic review","volume":"264","author":"Evers","year":"2017","journal-title":"J. Neurol."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"8822598","DOI":"10.1155\/2020\/8822598","article-title":"IoT-Based Healthcare Support System for Alzheimer\u2019s Patients","volume":"2020","author":"Oskouei","year":"2020","journal-title":"Wirel. Commun. Mob. Comput."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"193060","DOI":"10.1109\/ACCESS.2020.3032497","article-title":"A Wearable Fall Detection System Based on Body Area Networks","volume":"8","author":"Blunda","year":"2020","journal-title":"IEEE Access"},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Feng, M., and Meunier, J. (2022). Skeleton Graph-Neural-Network-Based Human Action Recognition: A Survey. Sensors, 22.","DOI":"10.3390\/s22062091"},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Yin, J., Han, J., Wang, C., Zhang, B., and Zeng, X. (2019, January 17\u201319). A Skeleton-based Action Recognition System for Medical Condition Detection. Proceedings of the IEEE Biomedical Circuits and Systems Conference (BioCAS), Nara, Japan.","DOI":"10.1109\/BIOCAS.2019.8919127"},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Jiao, L., Chen, J., Liu, F., Yang, S., You, C., Liu, X., Li, L., and Hou, B. (2022). Graph Representation Learning Meets Computer Vision: A Survey. IEEE Trans. Artif. Intell., 1\u201322.","DOI":"10.1109\/TAI.2022.3194869"},{"key":"ref_12","unstructured":"Li, B., Dai, Y., Cheng, X., Chen, H., Lin, Y., and He, M. (2017, January 10\u201314). Skeleton based action recognition using translation-scale invariant image mapping and multi-scale deep CNN. Proceedings of the IEEE International Conference on Multimedia Expo Workshops (ICMEW), Hong Kong, China."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Shahroudy, A., Liu, J., Ng, T.T., and Wang, G. (2016, January 27\u201330). Ntu rgb+ d: A large scale dataset for 3d human activity analysis. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.115"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"1963","DOI":"10.1109\/TPAMI.2019.2896631","article-title":"View Adaptive Neural Networks for High Performance Skeleton-Based Human Action Recognition","volume":"41","author":"Zhang","year":"2019","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"709","DOI":"10.1049\/cvi2.12119","article-title":"A robust and efficient method for skeleton-based human action recognition and its application for cross-dataset evaluation","volume":"16","author":"Nguyen","year":"2018","journal-title":"IET Comput. Vis."},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"346","DOI":"10.1016\/j.patcog.2017.02.030","article-title":"Enhanced skeleton visualization for view invariant human action recognition","volume":"68","author":"Liu","year":"2017","journal-title":"Pattern Recognit."},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Monti, F., Boscaini, D., Masci, J., Rodola, E., Svoboda, J., and Bronstein, M.M. (2017, January 21\u201326). Geometric deep learning on graphs and manifolds using mixture model cnns. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.576"},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"161374","DOI":"10.1109\/ACCESS.2021.3131809","article-title":"Lightweight Long and Short-Range Spatial-Temporal Graph Convolutional Network for Skeleton-Based Action Recognition","volume":"9","author":"Chen","year":"2021","journal-title":"IEEE Access"},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Yan, S., Xiong, Y., and Lin, D. (2018, January 2\u20137). Spatial temporal graph convolutional networks for skeleton-based action recognition. Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA.","DOI":"10.1609\/aaai.v32i1.12328"},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Shi, L., Zhang, Y., Cheng, J., and Lu, H. (2019, January 16\u201320). Two-stream adaptive graph convolutional networks for skeleton-based action recognition. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.01230"},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"257","DOI":"10.1109\/34.910878","article-title":"The recognition of human movement using temporal templates","volume":"23","author":"Bobick","year":"2001","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Klaser, A., Marsza\u0142ek, M., and Schmid, C. (2008, January 1\u20134). A spatio-temporal descriptor based on 3d-gradients. Proceedings of the BMVC 2008-19th British Machine Vision Conference, Leeds, UK.","DOI":"10.5244\/C.22.99"},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.cviu.2014.01.002","article-title":"Action recognition using global spatio-temporal features derived from sparse representations","volume":"123","author":"Somasundaram","year":"2014","journal-title":"Comput. Vis. Image Underst."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Carreira, J., and Zisserman, A. (2017, January 21\u201326). Quo vadis, action recognition? a new model and the kinetics dataset. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.502"},{"key":"ref_25","unstructured":"Simonyan, K., and Zisserman, A. (2014, January 8\u201313). Two-Stream Convolutional Networks for Action Recognition in Videos. Proceedings of the 27th International Conference on Neural Information Processing Systems-Volume 1, NIPS\u201914, Montreal, QC, Canada."},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Vemulapalli, R., Arrate, F., and Chellappa, R. (2014, January 23\u201328). Human Action Recognition by Representing 3D Skeletons as Points in a Lie Group. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA.","DOI":"10.1109\/CVPR.2014.82"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Zhang, P., Lan, C., Xing, J., Zeng, W., Xue, J., and Zheng, N. (2017, January 22\u201329). View Adaptive Recurrent Neural Networks for High Performance Human Action Recognition From Skeleton Data. Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy.","DOI":"10.1109\/ICCV.2017.233"},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Li, S., Li, W., Cook, C., Zhu, C., and Gao, Y. (2018, January 18\u201322). Independently recurrent neural network (indrnn): Building a longer and deeper rnn. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00572"},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Wang, H., and Wang, L. (2017, January 21\u201326). Modeling temporal dynamics and spatial configurations of actions using two-stream recurrent neural networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.387"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Tu, J., Liu, M., and Liu, H. (2018, January 23\u201327). Skeleton-Based Human Action Recognition Using Spatial Temporal 3D Convolutional Neural Networks. Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), San Diego, CA, USA.","DOI":"10.1109\/ICME.2018.8486566"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Ke, Q., Bennamoun, M., An, S., Sohel, F., and Boussaid, F. (2017, January 21\u201326). A new representation of skeleton sequences for 3d action recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.486"},{"key":"ref_32","unstructured":"Niepert, M., Ahmed, M., and Kutzkov, K. (2016, January 20\u201322). Learning convolutional neural networks for graphs. Proceedings of the International Conference on Machine Learning, New York, NY, USA."},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"16","DOI":"10.1049\/iet-cvi.2017.0062","article-title":"Skeleton-based human activity recognition for elderly monitoring systems","volume":"12","author":"Hbali","year":"2018","journal-title":"IET Comput. Vis."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Gul, M.A., Yousaf, M.H., Nawaz, S., Ur Rehman, Z., and Kim, H. (2020). Patient Monitoring by Abnormal Human Activity Recognition Based on CNN Architecture. Electronics, 9.","DOI":"10.3390\/electronics9121993"},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"52277","DOI":"10.1109\/ACCESS.2018.2869790","article-title":"Human Action Monitoring for Healthcare Based on Deep Learning","volume":"6","author":"Gao","year":"2018","journal-title":"IEEE Access"},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Leibe, B., Matas, J., Sebe, N., and Welling, M. (2016, January 11\u201314). Spatio-Temporal LSTM with Trust Gates for 3D Human Action Recognition. Proceedings of the Computer Vision\u2014ECCV 2016, Amsterdam, The Netherlands.","DOI":"10.1007\/978-3-319-46454-1"},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Song, S., Lan, C., Xing, J., Zeng, W., and Liu, J. (2017, January 4\u20139). An End-to-End Spatio-Temporal Attention Model for Human Action Recognition from Skeleton Data. Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA.","DOI":"10.1609\/aaai.v31i1.11212"},{"key":"ref_38","doi-asserted-by":"crossref","unstructured":"Zheng, W., Li, L., Zhang, Z., Huang, Y., and Wang, L. (2019, January 8\u201312). Relational Network for Skeleton-Based Action Recognition. Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), Shanghai, China.","DOI":"10.1109\/ICME.2019.00147"},{"key":"ref_39","unstructured":"Li, C., Zhong, Q., Xie, D., and Pu, S. (2017, January 10\u201314). Skeleton-based action recognition with convolutional neural networks. Proceedings of the IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Hong Kong, China."},{"key":"ref_40","first-page":"11045","article-title":"Part-Level Graph Convolutional Network for Skeleton-Based Action Recognition","volume":"34","author":"Huang","year":"2020","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"ref_41","doi-asserted-by":"crossref","unstructured":"Shi, L., Zhang, Y., Cheng, J., and Lu, H. (2019, January 15\u201320). Skeleton-Based Action Recognition With Directed Graph Neural Networks. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00810"}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/22\/21\/8438\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T01:09:36Z","timestamp":1760144976000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/22\/21\/8438"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,11,2]]},"references-count":41,"journal-issue":{"issue":"21","published-online":{"date-parts":[[2022,11]]}},"alternative-id":["s22218438"],"URL":"https:\/\/doi.org\/10.3390\/s22218438","relation":{},"ISSN":["1424-8220"],"issn-type":[{"value":"1424-8220","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,11,2]]}}}