{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,24]],"date-time":"2026-03-24T22:45:23Z","timestamp":1774392323081,"version":"3.50.1"},"reference-count":22,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2025,1,31]],"date-time":"2025-01-31T00:00:00Z","timestamp":1738281600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,31]],"date-time":"2025-01-31T00:00:00Z","timestamp":1738281600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Guizhou Provincial Science and Technology Projects and Science and Technology Foundation of Guizhou Province","award":["QKH-ZK[2023]-015 and [2020]1Y262"],"award-info":[{"award-number":["QKH-ZK[2023]-015 and [2020]1Y262"]}]},{"name":"Guizhou Provincial Science and Technology Projects and Science and Technology Foundation of Guizhou Province","award":["QKH-ZK[2023]-015 and [2020]1Y262"],"award-info":[{"award-number":["QKH-ZK[2023]-015 and [2020]1Y262"]}]},{"name":"Guizhou Provincial Science and Technology Projects and Science and Technology Foundation of Guizhou Province","award":["QKH-ZK[2023]-015 and [2020]1Y262"],"award-info":[{"award-number":["QKH-ZK[2023]-015 and [2020]1Y262"]}]},{"name":"Guizhou Provincial Science and Technology Projects and Science and Technology Foundation of Guizhou Province","award":["QKH-ZK[2023]-015 and [2020]1Y262"],"award-info":[{"award-number":["QKH-ZK[2023]-015 and [2020]1Y262"]}]},{"name":"Guizhou Provincial Science and Technology Projects and Science and Technology Foundation of Guizhou Province","award":["QKH-ZK[2023]-015 and [2020]1Y262"],"award-info":[{"award-number":["QKH-ZK[2023]-015 and [2020]1Y262"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2025,3]]},"DOI":"10.1007\/s11760-025-03844-y","type":"journal-article","created":{"date-parts":[[2025,1,31]],"date-time":"2025-01-31T12:48:16Z","timestamp":1738327696000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Research on original environment folk dance movement evaluation based on spatio-temporal graph convolutional networks"],"prefix":"10.1007","volume":"19","author":[{"given":"Ziheng","family":"Miao","sequence":"first","affiliation":[]},{"given":"Weixing","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Jingwen","family":"Xie","sequence":"additional","affiliation":[]},{"given":"Liandan","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Ningfeng","family":"Hu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,1,31]]},"reference":[{"key":"3844_CR1","unstructured":"Guo, C.: Excavating original folk dances and preserving the charm of ecological arts. Art Panorama 19, 103\u2013105 ((2023))"},{"key":"3844_CR2","doi-asserted-by":"publisher","DOI":"10.1155\/2022\/1455849","volume":"2022","author":"Y Peng","year":"2022","unstructured":"Peng, Y.: Research on dance teaching based on motion capture system. Math. Probl. Eng. 2022, e1455849 (2022). https:\/\/doi.org\/10.1155\/2022\/1455849","journal-title":"Math. Probl. Eng."},{"key":"3844_CR3","doi-asserted-by":"publisher","first-page":"603","DOI":"10.1007\/s00371-020-01826-4","volume":"37","author":"K Wang","year":"2021","unstructured":"Wang, K., Zhang, G., Yang, J., Bao, H.: Dynamic human body re-construction and motion tracking with low-cost depth cameras. Vis. Comput. 37, 603\u2013618 (2021). https:\/\/doi.org\/10.1007\/s00371-020-01826-4","journal-title":"Vis. Comput."},{"key":"3844_CR4","doi-asserted-by":"publisher","first-page":"621","DOI":"10.1007\/s00371-019-01644-3","volume":"36","author":"Y Qin","year":"2020","unstructured":"Qin, Y., Mo, L., Li, C., Luo, J.: Skeleton-based action recognition by part-aware graph convolutional networks. Vis. Comput. 36, 621\u2013631 (2020). https:\/\/doi.org\/10.1007\/s00371-019-01644-3","journal-title":"Vis. Comput."},{"key":"3844_CR5","doi-asserted-by":"crossref","unstructured":"Moreno-Noguer, F.: 3D human pose estimation from a single image via distance matrix regression. In; Presented at the Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2017)","DOI":"10.1109\/CVPR.2017.170"},{"key":"3844_CR6","doi-asserted-by":"crossref","unstructured":"Pavlakos, G., Zhou, X., Derpanis, K.G., Daniilidis, K.: Coarse-To-fine volumetric prediction for single-image 3D human pose. In: Presented at the Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2017)","DOI":"10.1109\/CVPR.2017.139"},{"key":"3844_CR7","doi-asserted-by":"publisher","first-page":"1401","DOI":"10.1007\/s00371-019-01740-4","volume":"36","author":"J Wu","year":"2020","unstructured":"Wu, J., Hu, D., Xiang, F., Yuan, X., Su, J.: 3D human pose estimation by depth map. Vis. Comput. 36, 1401\u20131410 (2020). https:\/\/doi.org\/10.1007\/s00371-019-01740-4","journal-title":"Vis. Comput."},{"key":"3844_CR8","doi-asserted-by":"publisher","first-page":"1062","DOI":"10.1016\/j.gaitpost.2014.01.008","volume":"39","author":"B Galna","year":"2014","unstructured":"Galna, B., Barry, G., Jackson, D., Mhiripiri, D., Olivier, P., Rochester, L.: accuracy of the microsoft kinect sensor for measuring movement in people with parkinson\u2019s disease. Gait Posture 39, 1062\u20131068 (2014). https:\/\/doi.org\/10.1016\/j.gaitpost.2014.01.008","journal-title":"Gait Posture"},{"key":"3844_CR9","doi-asserted-by":"publisher","first-page":"136","DOI":"10.1016\/j.cviu.2015.11.016","volume":"148","author":"L Tao","year":"2016","unstructured":"Tao, L., Paiement, A., Damen, D., Mirmehdi, M., Hannuna, S., Camplani, M., Burghardt, T., Craddock, I.: A comparative study of pose representation and dynamics modelling for online motion quality assessment. Comput. Vis. Image Underst. 148, 136\u2013152 (2016). https:\/\/doi.org\/10.1016\/j.cviu.2015.11.016","journal-title":"Comput. Vis. Image Underst."},{"key":"3844_CR10","doi-asserted-by":"publisher","unstructured":"Pirsiavash, H., Vondrick, C., Torralba, A.: Assessing the quality of actions. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds) Computer Vision\u2014ECCV 2014, pp 556\u2013571. Springer International Publishing, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10599-4_36","DOI":"10.1007\/978-3-319-10599-4_36"},{"key":"3844_CR11","doi-asserted-by":"publisher","unstructured":"Parmar, P., Morris, B.T.: Measuring the quality of exercises. In: 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC). pp. 2241\u20132244 (2016). https:\/\/doi.org\/10.1109\/EMBC.2016.7591175.","DOI":"10.1109\/EMBC.2016.7591175"},{"key":"3844_CR12","doi-asserted-by":"publisher","unstructured":"Antunes, M., Baptista, R., Demisse, G., Aouada, D., Ottersten, B.: Visual and human-interpretable feedback for assisting physical activity. In: Hua, G., J\u00e9gou, H. (eds.) Computer Vision\u2014ECCV 2016 Workshops, pp. 115\u2013129. Springer International Publishing, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-48881-3_9","DOI":"10.1007\/978-3-319-48881-3_9"},{"key":"3844_CR13","doi-asserted-by":"publisher","unstructured":"Home self-training: visual feedback for assisting physical activity for stroke survivors. Computer Methods and Programs in Biomedicine. 176, 111\u2013120 (2019). https:\/\/doi.org\/10.1016\/j.cmpb.2019.04.019.","DOI":"10.1016\/j.cmpb.2019.04.019"},{"key":"3844_CR14","unstructured":"Deng, X.: Development golf swing analysis system with auxiliary training based on Kinect motioncapture. Beijing University of Posts and Telecommunications, (2013)"},{"key":"3844_CR15","unstructured":"Li, W.: Cultural ecology and cultural inheritance of the Tujia Pendulum Dance. J. Central South Univ. National. Humanit. Soc. Sci. 27, 143\u2013146 (2007)."},{"key":"3844_CR16","doi-asserted-by":"publisher","unstructured":"Liu, C.: Studies on the aesthetic paradigm of the Bouyei dance. Guizhou Ethnic Studies. 35, 79\u201382 (2014). https:\/\/doi.org\/10.13965\/j.cnki.gzmzyj10026959.2014.12.020","DOI":"10.13965\/j.cnki.gzmzyj10026959.2014.12.020"},{"key":"3844_CR17","unstructured":"Guan, Y.: The \"chaff bag dance\" of the Buyi. Sports Culture Guide. 77(2), 1 (1990)"},{"key":"3844_CR18","doi-asserted-by":"publisher","unstructured":"Wei, Y.: Research on the inheritance and development of Guizhou Buyei bamboo drum dance. Contemp. Sports Sci. Technol. 6(30), 180\u2013182 (2016). https:\/\/doi.org\/10.16655\/j.cnki.2095-2813.2016.30.180","DOI":"10.16655\/j.cnki.2095-2813.2016.30.180"},{"key":"3844_CR19","unstructured":"Liu, Y, & Dong, Y: On performance and cultural value of Miao's panpipe dance in Guizhou Province. J. Beijing Dance Acad., 4, 85\u201388 (2011)"},{"key":"3844_CR20","doi-asserted-by":"publisher","unstructured":"Yan, S., Xiong, Y., Lin, D.: Spatial temporal graph convolutional networks for skeleton-based action recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence. 32, (2018). https:\/\/doi.org\/10.1609\/aaai.v32i1.12328.","DOI":"10.1609\/aaai.v32i1.12328"},{"key":"3844_CR21","doi-asserted-by":"publisher","first-page":"2684","DOI":"10.1364\/OL.484455","volume":"48","author":"H Xu","year":"2023","unstructured":"Xu, H., Song, X.-K., Ye, L., Wang, D.: Efficient and robust chiral discrimination by invariant-based inverse engineering. Opt. Lett. 48, 2684\u20132687 (2023). https:\/\/doi.org\/10.1364\/OL.484455","journal-title":"Opt. Lett."},{"key":"3844_CR22","doi-asserted-by":"publisher","first-page":"2","DOI":"10.3390\/data3010002","volume":"3","author":"A Vakanski","year":"2018","unstructured":"Vakanski, A., Jun, H., Paul, D., Baker, R.: A data set of human body movements for physical rehabilitation exercises. Data. 3, 2 (2018). https:\/\/doi.org\/10.3390\/data3010002","journal-title":"Data."}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-03844-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-025-03844-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-03844-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,13]],"date-time":"2025-02-13T14:51:06Z","timestamp":1739458266000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-025-03844-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1,31]]},"references-count":22,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2025,3]]}},"alternative-id":["3844"],"URL":"https:\/\/doi.org\/10.1007\/s11760-025-03844-y","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"value":"1863-1703","type":"print"},{"value":"1863-1711","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,1,31]]},"assertion":[{"value":"13 October 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 December 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 January 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 January 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no relevant financial or nonfinancial interests to disclose.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"266"}}