{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T18:59:07Z","timestamp":1775242747953,"version":"3.50.1"},"reference-count":38,"publisher":"MDPI AG","issue":"14","license":[{"start":{"date-parts":[[2022,7,20]],"date-time":"2022-07-20T00:00:00Z","timestamp":1658275200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"<jats:p>Various genres of dance, such as Yosakoi Soran, have contributed to the health of many people and contributed to their sense of belonging to a community. However, due to the effects of COVID-19, various face-to-face activities have been restricted and group dance practice has become difficult. Hence, there is a need to facilitate remote dance practice. In this paper, we propose a system for detecting and visualizing the very important dance motions known as stops. We measure dance movements by motion capture and calculate the features of each movement based on velocity and acceleration. Using a neural network to learn motion features, the system detects stops and visualizes them using a human-like 3D model. In an experiment using dance data, the proposed method obtained highly accurate stop detection results and demonstrated its effectiveness as an information and communication technology support for remote group dance practice.<\/jats:p>","DOI":"10.3390\/s22145402","type":"journal-article","created":{"date-parts":[[2022,7,21]],"date-time":"2022-07-21T03:34:40Z","timestamp":1658374480000},"page":"5402","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Detecting and Visualizing Stops in Dance Training by Neural Network Based on Velocity and Acceleration"],"prefix":"10.3390","volume":"22","author":[{"given":"Yuuki","family":"Jin","sequence":"first","affiliation":[{"name":"Division of Information and Electronic Engineering, Muroran Institute of Technology, Muroran 050-8585, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0881-7544","authenticated-orcid":false,"given":"Genki","family":"Suzuki","sequence":"additional","affiliation":[{"name":"Division of Information and Electronic Engineering, Muroran Institute of Technology, Muroran 050-8585, Japan"}]},{"given":"Hiroyuki","family":"Shioya","sequence":"additional","affiliation":[{"name":"Division of Information and Electronic Engineering, Muroran Institute of Technology, Muroran 050-8585, Japan"}]}],"member":"1968","published-online":{"date-parts":[[2022,7,20]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"41","DOI":"10.1080\/14647890120058302","article-title":"The relationship between play and dance","volume":"2","author":"Lindqvist","year":"2001","journal-title":"Res. Danc. Educ."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"155","DOI":"10.1177\/1084822310384689","article-title":"The health benefits of dance","volume":"23","author":"Alpert","year":"2011","journal-title":"Home Health Care Manag. Pract."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"33","DOI":"10.1080\/07303084.2008.10598161","article-title":"Health and the power of dance","volume":"79","author":"Ward","year":"2008","journal-title":"J. Phys. Educ. Recreat. Danc."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"157","DOI":"10.1080\/14647893.2011.579596","article-title":"The teaching artist: A model for university dance teacher training","volume":"12","author":"Huddy","year":"2011","journal-title":"Res. Danc. Educ."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"199","DOI":"10.1080\/14647893.2018.1476479","article-title":"Creating a cultural dance community of practice: Building authentic Latin American dance experiences","volume":"19","author":"Pedro","year":"2018","journal-title":"Res. Danc. Educ."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"53","DOI":"10.1080\/14647890050006587","article-title":"Power, service, and reflexivity in a community dance project","volume":"1","author":"Green","year":"2000","journal-title":"Res. Danc. Educ."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"353","DOI":"10.1080\/19325037.2008.10599062","article-title":"Cultural dance and health: A review of the literature","volume":"39","author":"Olvera","year":"2008","journal-title":"Am. J. Health Educ."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"21","DOI":"10.2307\/1478188","article-title":"Performance, transformation, and community: Contra dance in New England","volume":"25","author":"Hast","year":"1993","journal-title":"Danc. Res. J."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"237","DOI":"10.1080\/2052546.2003.11949263","article-title":"The opposite of powwow: Ignoring and incorporating the intertribal war dance in the Oklahoma stomp dance community","volume":"48","author":"Jackson","year":"2003","journal-title":"Plains Anthropol."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"36","DOI":"10.1177\/016235320402800103","article-title":"The dance teacher: The ideal case and daily reality","volume":"28","year":"2004","journal-title":"J. Educ. Gift."},{"key":"ref_11","first-page":"7","article-title":"Acceptance of YouTube applied to dance learning","volume":"10","author":"Hong","year":"2020","journal-title":"Int. J. Inf. Educ. Technol."},{"key":"ref_12","unstructured":"Akiba, F. (2017, January 26\u201330). YOSAKOI SORAN as a site of re-localization and its relationship to Japanese pop culture. Proceedings of the 13th World Congress of the International Association for Semiotic Studies (IASS\/AIS), Kaunas, Lithuania."},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"187","DOI":"10.1109\/TLT.2010.27","article-title":"A virtual reality dance training system using motion capture technology","volume":"4","author":"Chan","year":"2011","journal-title":"IEEE Trans. Learn. Technol."},{"key":"ref_14","unstructured":"Hachimura, K., Kato, H., and Tamura, H. (2004, January 22\u201324). A prototype dance training support system with motion capture and mixed reality technologies. Proceedings of the 13th IEEE International Workshop on Robot and Human Interactive Communication, Kurashiki, Japan."},{"key":"ref_15","unstructured":"Shiratori, T., Nakazawa, A., and Ikeuchi, K. (2004, January 17\u201319). Detecting dance motion structure using motion capture and musical information. Proceedings of the 10th International Conference on Virtual Systems and Multimedia, Ogaki, Japan."},{"key":"ref_16","unstructured":"Nakazawa, A., Nakaoka, S., Ikeuchi, K., and Yokoi, K. (October, January 30). Imitating human dance motions through motion structure analysis. Proceedings of the IEEE\/RSJ International Conference on Intelligent Robots and Systems, Lausanne, Switzerland."},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"71","DOI":"10.1002\/scj.20250","article-title":"Analysis of Japanese dance movements using motion capture system","volume":"37","author":"Yoshimura","year":"2006","journal-title":"Syst. Comput. Jpn."},{"key":"ref_18","unstructured":"Solberg, R.T., and Jensenius, A.R. (2016, January 9\u201312). Optical or inertial? Evaluation of two motion capture systems for studies of dancing to electronic dance music. Proceedings of the Systems, Man, and Cybernetics Conferences, Budapest, Hungary."},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Camurri, A., El Raheb, K., Even-Zohar, O., Ioannidis, Y., Markatzi, A., Matos, J.M., Morley-Fletcher, E., Palacio, P., Romero, M., and Sarti, A. (2016, January 5\u20136). WhoLoDancE: Towards a methodology for selecting motion capture data across different dance learning practice. Proceedings of the 3rd International Symposium on Movement and Computing, Thessaloniki, Greece.","DOI":"10.1145\/2948910.2948912"},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/2755566","article-title":"Folk dance evaluation using laban movement analysis","volume":"8","author":"Aristidou","year":"2015","journal-title":"ACM J. Comput. Cult. Herit."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"100974","DOI":"10.1016\/j.tsc.2021.100974","article-title":"Modern social dance teaching approaches: Studying creative and communicative components","volume":"43","author":"Wang","year":"2022","journal-title":"Think. Ski. Creat."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"612","DOI":"10.1016\/j.patcog.2017.12.007","article-title":"Motion analysis: Action detection, recognition and evaluation based on motion capture data","volume":"76","author":"Patrona","year":"2018","journal-title":"Pattern Recognit."},{"key":"ref_23","unstructured":"Zhang, Z., Li, J., Manikopoulos, C., Jorgenson, J., and Ucles, J. (2001, January 21\u201323). HIDE: A hierarchical network intrusion detection system using statistical preprocessing and neural network classification. Proceedings of the IEEE Workshop on Information Assurance and Security, St. Petersburg, Russia."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Kim, H.S., Hong, N., Kim, M., Yoon, S.G., Yu, H.W., Kong, H.J., Kim, S.J., Chai, Y.J., Choi, H.J., and Choi, J.Y. (2019). Application of a perception neuron\u00ae system in simulation-based surgical training. J. Clin. Med., 8.","DOI":"10.3390\/jcm8010124"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Akamatsu, Y., Maeda, K., Ogawa, T., and Haseyama, M. (2021, January 6\u201311). Classification of expert-novice level using eye tracking and motion data via conditional multimodal variational autoencoder. Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada.","DOI":"10.1109\/ICASSP39728.2021.9414361"},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"1509","DOI":"10.1080\/01621459.1970.10481180","article-title":"Distribution of residual autocorrelations in autoregressive-integrated moving average time series models","volume":"65","author":"Box","year":"1970","journal-title":"J. Am. Stat. Assoc."},{"key":"ref_27","unstructured":"Agarap, A.F. (2018). Deep learning using rectified linear units (ReLU). arXiv."},{"key":"ref_28","first-page":"1603","article-title":"Gated softmax classification","volume":"23","author":"Memisevic","year":"2010","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"310","DOI":"10.1007\/s00521-005-0467-y","article-title":"Revisiting squared-error and cross-entropy functions for training neural network classifiers","volume":"14","author":"Kline","year":"2005","journal-title":"Neural Comput. Appl."},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Lin, C.Y., Yang, Z.H., Zhou, H.W., Yang, T.N., Chen, H.N., and Shih, T.K. (2018, January 10\u201312). Combining leap motion with unity for virtual glove puppets. Proceedings of the IEEE International Conference on Artificial Intelligence and Virtual Reality (AIVR), Taichung, Taiwan.","DOI":"10.1109\/AIVR.2018.00059"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Guo, H., and Sung, Y. (2020). Movement estimation using soft sensors based on Bi-LSTM and two-layer LSTM for human motion capture. Sensors, 20.","DOI":"10.3390\/s20061801"},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Siami-Namini, S., Tavakoli, N., and Namin, A.S. (2018, January 17\u201320). A comparison of ARIMA and LSTM in forecasting time series. Proceedings of the 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL, USA.","DOI":"10.1109\/ICMLA.2018.00227"},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"273","DOI":"10.1007\/BF00994018","article-title":"Support-vector networks","volume":"20","author":"Cortes","year":"1995","journal-title":"Mach. Learn."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"21","DOI":"10.1109\/TIT.1967.1053964","article-title":"Nearest neighbor pattern classification","volume":"13","author":"Cover","year":"1967","journal-title":"IEEE Trans. Inf. Theory"},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"5","DOI":"10.1023\/A:1010933404324","article-title":"Random forests","volume":"45","author":"Breiman","year":"2001","journal-title":"Mach. Learn."},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"172","DOI":"10.1109\/TPAMI.2019.2929257","article-title":"OpenPose: Realtime multi-person 2D pose estimation using part affinity fields","volume":"43","author":"Cao","year":"2019","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Chen, C.H., and Ramanan, D. (2017, January 21\u201326). 3d human pose estimation= 2d pose estimation+ matching. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.610"},{"key":"ref_38","unstructured":"Wang, L., Chen, Y., Guo, Z., Qian, K., Lin, M., Li, H., and Ren, J.S. (November, January 27). Generalizing monocular 3d human pose estimation in the wild. Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV) Workshops, Seoul, Korea."}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/22\/14\/5402\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T23:54:16Z","timestamp":1760140456000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/22\/14\/5402"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,7,20]]},"references-count":38,"journal-issue":{"issue":"14","published-online":{"date-parts":[[2022,7]]}},"alternative-id":["s22145402"],"URL":"https:\/\/doi.org\/10.3390\/s22145402","relation":{},"ISSN":["1424-8220"],"issn-type":[{"value":"1424-8220","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,7,20]]}}}