{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,16]],"date-time":"2026-03-16T21:35:48Z","timestamp":1773696948502,"version":"3.50.1"},"reference-count":44,"publisher":"MDPI AG","issue":"17","license":[{"start":{"date-parts":[[2020,8,28]],"date-time":"2020-08-28T00:00:00Z","timestamp":1598572800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100002641","name":"Konkuk University","doi-asserted-by":"publisher","award":["2017"],"award-info":[{"award-number":["2017"]}],"id":[{"id":"10.13039\/501100002641","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"<jats:p>In taekwondo, poomsae (i.e., form) competitions have no quantitative scoring standards, unlike gyeorugi (i.e., full-contact sparring) in the Olympics. Consequently, there are diverse fairness issues regarding poomsae evaluation, and the demand for quantitative evaluation tools is increasing. Action recognition is a promising approach, but the extreme and rapid actions of taekwondo complicate its application. This study established the Taekwondo Unit technique Human Action Dataset (TUHAD), which consists of multimodal image sequences of poomsae actions. TUHAD contains 1936 action samples of eight unit techniques performed by 10 experts and captured by two camera views. A key frame-based convolutional neural network architecture was developed for taekwondo action recognition, and its accuracy was validated for various input configurations. A correlation analysis of the input configuration and accuracy demonstrated that the proposed model achieved a recognition accuracy of up to 95.833% (lowest accuracy of 74.49%). This study contributes to the research and development of taekwondo action recognition.<\/jats:p>","DOI":"10.3390\/s20174871","type":"journal-article","created":{"date-parts":[[2020,8,28]],"date-time":"2020-08-28T09:17:08Z","timestamp":1598606228000},"page":"4871","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":27,"title":["TUHAD: Taekwondo Unit Technique Human Action Dataset with Key Frame-Based CNN Action Recognition"],"prefix":"10.3390","volume":"20","author":[{"given":"Jinkue","family":"Lee","sequence":"first","affiliation":[{"name":"Department of Mechanical Engineering, Konkuk University, 120 Neungdong-ro, Jayang-dong, Gwangjin-gu, Seoul 05029, Korea"}]},{"given":"Hoeryong","family":"Jung","sequence":"additional","affiliation":[{"name":"Department of Mechanical Engineering, Konkuk University, 120 Neungdong-ro, Jayang-dong, Gwangjin-gu, Seoul 05029, Korea"}]}],"member":"1968","published-online":{"date-parts":[[2020,8,28]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Wei, H., Chopada, P., and Kehtarnavaz, N. (2020). C-MHAD: Continuous Multimodal Human Action Dataset of Simultaneous Video and Inertial Sensing. Sensors, 20.","DOI":"10.3390\/s20102905"},{"key":"ref_2","unstructured":"Ren, H., and Xu, G. (2002, January 21\u201321). Human action recognition in smart classroom. Proceedings of the Fifth IEEE International Conference on Automatic Face Gesture Recognition, Washington, DC, USA."},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Rautaray, S.S., and Agrawal, A. (2011, January 17\u201319). Interaction with virtual game through hand gesture recognition. Proceedings of the 2011 International Conference on Multimedia, Signal Processing and Communication Technologies, Aligarh, India.","DOI":"10.1109\/MSPCT.2011.6150485"},{"key":"ref_4","unstructured":"Kong, Y., Zhang, X., Wei, Q., Hu, W., and Jia, Y. (2008, January 8\u201311). Group action recognition in soccer videos. Proceedings of the 2008 19th International Conference on Pattern Recognition, Tampa, FL, USA."},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Zhang, L., Hsieh, J.-C., Ting, T.-T., Huang, Y.-C., Ho, Y.-C., and Ku, L.-K. (2012, January 16\u201318). A Kinect based Golf Swing Score and Grade System using GMM and SVM. Proceedings of the 2012 5th International Congress on Image and Signal Processing, Chongqing, China.","DOI":"10.1109\/CISP.2012.6469827"},{"key":"ref_6","unstructured":"Zhang, L., Hsieh, J.C., and Wang, J. (2012, January 24\u201326). A Kinect-based golf swing classification system using HMM and Neuro-Fuzzy. Proceedings of the 2012 International Conference on Computer Science and Information Processing (CSIP), Xian, China."},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Zhu, G., Xu, C., Huang, Q., Gao, W., and Xing, L. (2006, January 12\u201316). Player action recognition in broadcast tennis video with applications to semantic analysis of sports game. Proceedings of the 14th Annual ACM International Conference on Multimedia\u2014MULTIMEDIA\u201906, Santa Barbara, CA, USA.","DOI":"10.1145\/1180639.1180728"},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"FarajiDavar, N., de Campos, T., Kittler, J., and Yan, F. (2011, January 6\u201313). Transductive transfer learning for action recognition in tennis games. Proceedings of the 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), Barcelona, Spain.","DOI":"10.1109\/ICCVW.2011.6130434"},{"key":"ref_9","unstructured":"Zhu, G., Xu, C., Huang, Q., and Gao, W. (2006, January 20\u201324). Action Recognition in Broadcast Tennis Video. Proceedings of the 18th International Conference on Pattern Recognition (ICPR\u201906), Hong Kong, China."},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Martin, P.-E., Benois-Pineau, J., Peteri, R., and Morlier, J. (2018, January 4\u20136). Sport Action Recognition with Siamese Spatio-Temporal CNNs: Application to Table Tennis. Proceedings of the 2018 International Conference on Content-Based Multimedia Indexing (CBMI), La Rochelle, France.","DOI":"10.1109\/CBMI.2018.8516488"},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Piergiovanni, A.J., and Ryoo, M.S. (2018). Fine-grained Activity Recognition in Baseball Videos. arXiv.","DOI":"10.1109\/CVPRW.2018.00226"},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Pham, H.H., Salmane, H., Khoudour, L., Crouzil, A., Velastin, S.A., and Zegers, P. (2020). A Unified Deep Framework for Joint 3D Pose Estimation and Action Recognition from a Single RGB Camera. Sensors, 20.","DOI":"10.3390\/s20071825"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Dong, J., Gao, Y., Lee, H.J., Zhou, H., Yao, Y., Fang, Z., and Huang, B. (2020). Action Recognition Based on the Fusion of Graph Convolutional Networks with High Order Features. Appl. Sci., 10.","DOI":"10.3390\/app10041482"},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Wang, H., Song, Z., Li, W., and Wang, P. (2020). A Hybrid Network for Large-Scale Action Recognition from RGB and Depth Modalities. Sensors, 20.","DOI":"10.3390\/s20113305"},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Du, Y., Fu, Y., and Wang, L. (2015, January 3\u20136). Skeleton based action recognition with convolutional neural network. Proceedings of the 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR), Kuala Lumpur, Malaysia.","DOI":"10.1109\/ACPR.2015.7486569"},{"key":"ref_16","unstructured":"Ravanbakhsh, M., Mousavi, H., Rastegari, M., Murino, V., and Davis, L.S. (2015). Action Recognition with Image Based CNN Features. arXiv."},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C., Pinz, A., and Zisserman, A. (2016). Convolutional Two-Stream Network Fusion for Video Action Recognition. arXiv.","DOI":"10.1109\/CVPR.2016.213"},{"key":"ref_18","unstructured":"Li, B., Dai, Y., Cheng, X., Chen, H., Lin, Y., and He, M. (2017, January 10\u201314). Skeleton based action recognition using translation-scale invariant image mapping and multi-scale deep CNN. Proceedings of the 2017 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Hong Kong, China."},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Ercolano, G., Riccio, D., and Rossi, S. (September, January 28). Two deep approaches for ADL recognition: A multi-scale LSTM and a CNN-LSTM with a 3D matrix skeleton representation. Proceedings of the 2017 26th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), Lisbon, Portugal.","DOI":"10.1109\/ROMAN.2017.8172406"},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Ke, Q., Bennamoun, M., An, S., Sohel, F., and Boussaid, F. (2017, January 21\u201326). A New Representation of Skeleton Sequences for 3D Action Recognition. Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.486"},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"22901","DOI":"10.1007\/s11042-018-5642-0","article-title":"3D skeleton based action recognition by video-domain translation-scale invariant mapping and multi-scale dilated CNN","volume":"77","author":"Li","year":"2018","journal-title":"Multimed. Tools Appl."},{"key":"ref_22","unstructured":"Ding, Z., Wang, P., Ogunbona, P.O., and Li, W. (2017, January 10\u201314). Investigation of different skeleton features for CNN-based 3D action recognition. Proceedings of the 2017 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Hong Kong, China."},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Liu, C., Hu, Y., Li, Y., Song, S., and Liu, J. (2017). PKU-MMD: A Large Scale Benchmark for Continuous Multi-Modal Human Action Understanding. arXiv.","DOI":"10.1145\/3132734.3132739"},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Shahroudy, A., Liu, J., Ng, T.-T., and Wang, G. (2016). NTU RGB+D: A Large Scale Dataset for 3D Human Activity Analysis. arXiv.","DOI":"10.1109\/CVPR.2016.115"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Liu, J., Shahroudy, A., Perez, M., Wang, G., Duan, L.-Y., and Kot, A.C. (2019). NTU RGB+D 120: A Large-Scale Benchmark for 3D Human Activity Understanding. IEEE Trans. Pattern Anal. Mach. Intell., 1.","DOI":"10.1109\/TPAMI.2019.2916873"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Chen, C., Jafari, R., and Kehtarnavaz, N. (2015, January 27\u201330). UTD-MHAD: A multimodal dataset for human action recognition utilizing a depth camera and a wearable inertial sensor. Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada.","DOI":"10.1109\/ICIP.2015.7350781"},{"key":"ref_27","unstructured":"Goma, J.C., Bustos, M.S., Sebastian, J.A., and Macrohon, J.J.E. (2019, January 9\u201311). Detection of Taekwondo Kicks Using RGB-D Sensors. Proceedings of the 2019 3rd International Conference on Software and e-Business, Tokyo, Japan."},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"1453","DOI":"10.1109\/TPAMI.2019.2898954","article-title":"Skeleton-Based Online Action Prediction Using Scale Selection Network","volume":"42","author":"Liu","year":"2020","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"3007","DOI":"10.1109\/TPAMI.2017.2771306","article-title":"Skeleton-Based Action Recognition Using Spatio-Temporal LSTM Network with Trust Gates","volume":"40","author":"Liu","year":"2018","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Livingston, M.A., Sebastian, J., Ai, Z., and Decker, J.W. (2012, January 4\u20138). Performance measurements for the Microsoft Kinect skeleton. Proceedings of the 2012 IEEE Virtual Reality (VR), Costa Mesa, CA, USA.","DOI":"10.1109\/VR.2012.6180911"},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"22","DOI":"10.1016\/j.imavis.2017.02.002","article-title":"Martial Arts, Dancing and Sports dataset: A challenging stereo and multi-view dataset for 3D human pose estimation","volume":"61","author":"Zhang","year":"2017","journal-title":"Image Vis. Comput."},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Moeslund, T.B., Thomas, G., and Hilton, A. (2014). Action Recognition in Realistic Sports Videos. Computer Vision in Sports, Springer International Publishing. Advances in Computer Vision and Pattern, Recognition.","DOI":"10.1007\/978-3-319-09396-3"},{"key":"ref_33","unstructured":"Soomro, K., Zamir, A.R., and Shah, M. (2012). UCF101: A Dataset of 101 Human Actions Classes from Videos in the Wild. arXiv."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Heinz, E.A., Kunze, K.S., Gruber, M., Bannach, D., and Lukowicz, P. (2006, January 22\u201324). Using Wearable Sensors for Real-Time Recognition Tasks in Games of Martial Arts\u2014An Initial Experiment. Proceedings of the 2006 IEEE Symposium on Computational Intelligence and Games, Reno, NV, USA.","DOI":"10.1109\/CIG.2006.311687"},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Salazar, K.A., Sibaja Garcia, J.E., Mateus, A.S., and Percybrooks, W.S. (2017, January 4\u20136). Autonomous recognition of martial arts forms using RGB-D cameras. Proceedings of the 2017 Congreso Internacional de Innovacion y Tendencias en Ingenieria (CONIITI), Bogota, Colombia.","DOI":"10.1109\/CONIITI.2017.8273323"},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Stasinopoulos, S., and Maragos, P. (October, January 30). Human action recognition using Histographic methods and hidden Markov models for visual martial arts applications. Proceedings of the 2012 19th IEEE International Conference on Image Processing, Orlando, FL, USA.","DOI":"10.1109\/ICIP.2012.6466967"},{"key":"ref_37","doi-asserted-by":"crossref","first-page":"13135","DOI":"10.1007\/s11042-015-2901-1","article-title":"Motion recognition technology based remote Taekwondo Poomsae evaluation system","volume":"75","author":"Choi","year":"2016","journal-title":"Multimed. Tools Appl."},{"key":"ref_38","unstructured":"Seo, J.M., Jang, I.K., Choi, J.H., and Lee, S.M. (2009, January 20\u201322). A Study of the Taekwondo Poomsae Recognition System Used by Motion Recognition Techniques. Proceedings of the 2009 International Conference on Multimedia Information Technology and Applications, Osaka, Japan."},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"13643","DOI":"10.1007\/s11042-017-4979-0","article-title":"Automatic analysis of complex athlete techniques in broadcast taekwondo video","volume":"77","author":"Kong","year":"2018","journal-title":"Multimed. Tools Appl."},{"key":"ref_40","unstructured":"Simonyan, K., and Zisserman, A. (2014). Two-Stream Convolutional Networks for Action Recognition in Videos. arXiv."},{"key":"ref_41","doi-asserted-by":"crossref","unstructured":"Zhang, B., Wang, L., Wang, Z., Qiao, Y., and Wang, H. (2016, January 27\u201330). Real-Time Action Recognition with Enhanced Motion Vector CNNs. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.297"},{"key":"ref_42","doi-asserted-by":"crossref","unstructured":"Dehzangi, O., Taherisadr, M., and ChangalVala, R. (2017). IMU-Based Gait Recognition Using Convolutional Neural Networks and Multi-Sensor Fusion. Sensors, 17.","DOI":"10.3390\/s17122735"},{"key":"ref_43","unstructured":"Kingma, D.P., and Ba, J. (2017). Adam: A Method for Stochastic Optimization. arXiv."},{"key":"ref_44","unstructured":"(2020, August 07). UCF Sports Action Data Set. Available online: https:\/\/www.crcv.ucf.edu\/data\/UCF_Sports_Action.php."}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/20\/17\/4871\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T10:04:13Z","timestamp":1760177053000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/20\/17\/4871"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,8,28]]},"references-count":44,"journal-issue":{"issue":"17","published-online":{"date-parts":[[2020,9]]}},"alternative-id":["s20174871"],"URL":"https:\/\/doi.org\/10.3390\/s20174871","relation":{},"ISSN":["1424-8220"],"issn-type":[{"value":"1424-8220","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,8,28]]}}}