{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T06:20:24Z","timestamp":1764829224491,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":35,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,5,24]],"date-time":"2024-05-24T00:00:00Z","timestamp":1716508800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"National Natural Science Foundation of China","award":["62162068, 62061049"],"award-info":[{"award-number":["62162068, 62061049"]}]},{"name":"Science Research Fund Project of Yunnan Provincial Department of Education","award":["2021Y027"],"award-info":[{"award-number":["2021Y027"]}]},{"name":"Yunnan Province Ten Thousand Talents Program and Yunling Scholars Special Project","award":["YNWR-YLXZ-2018-022"],"award-info":[{"award-number":["YNWR-YLXZ-2018-022"]}]},{"name":"Yunnan Provincial Science and Technology Department - Yunnan University Double First Class Construction Joint Fund","award":["202301BF070001-025"],"award-info":[{"award-number":["202301BF070001-025"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,5,24]]},"DOI":"10.1145\/3670105.3670151","type":"proceedings-article","created":{"date-parts":[[2024,7,29]],"date-time":"2024-07-29T18:29:36Z","timestamp":1722277776000},"page":"273-281","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["YNU-Dance: A Multimodal Ethnic Dance Action Dataset"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-2946-3777","authenticated-orcid":false,"given":"Shouzheng","family":"Sun","sequence":"first","affiliation":[{"name":"Yunnan University, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-9516-3832","authenticated-orcid":false,"given":"Qianhan","family":"Tang","sequence":"additional","affiliation":[{"name":"Yunnan University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2320-229X","authenticated-orcid":false,"given":"Yanan","family":"Liu","sequence":"additional","affiliation":[{"name":"Yunnan University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0404-6941","authenticated-orcid":false,"given":"Hao","family":"Zhang","sequence":"additional","affiliation":[{"name":"Yunnan University, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-9947-9886","authenticated-orcid":false,"given":"Qianyun","family":"Song","sequence":"additional","affiliation":[{"name":"Yunnan University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4602-3550","authenticated-orcid":false,"given":"Dan","family":"Xu","sequence":"additional","affiliation":[{"name":"Yunnan University, China"}]}],"member":"320","published-online":{"date-parts":[[2024,7,29]]},"reference":[{"unstructured":"Omid Alemi Jules Fran\u00e7oise and Philippe Pasquier. [n. d.]. GrooveNet: Real-Time Music-Driven Dance Movement Generation using Artificial Neural Networks *. ([n. d.]).","key":"e_1_3_2_1_1_1"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_2_1","DOI":"10.1109\/ICIP.2015.7350781"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_3_1","DOI":"10.1109\/iccv48922.2021.01311"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_4_1","DOI":"10.1109\/iccv.2015.368"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_5_1","DOI":"10.1109\/cvpr.2015.7298714"},{"key":"e_1_3_2_1_6_1","volume-title":"Dancing to Music","author":"Lee Hsin-Ying","year":"2019","unstructured":"Hsin-Ying Lee, Xiaodong Yang, Mingyu Li, Ting-Chun Wang, Yu-Ding Lu, Ming-Hsuan Yang, and Jan Kautz. 2019. Dancing to Music. Cornell University - arXiv,Cornell University - arXiv (Nov 2019)."},{"key":"e_1_3_2_1_7_1","volume-title":"Listen to Dance: Music-driven choreography generation using Autoregressive Encoder-Decoder Network","author":"Lee Juheon","year":"2018","unstructured":"Juheon Lee, Seohyun Kim, and Kyogu Lee. 2018. Listen to Dance: Music-driven choreography generation using Autoregressive Encoder-Decoder Network. Cornell University - arXiv,Cornell University - arXiv (Nov 2018)."},{"key":"e_1_3_2_1_8_1","volume-title":"Hierarchically Decomposed Graph Convolutional Networks for Skeleton-Based Action Recognition. (Aug","author":"Lee Jungho","year":"2022","unstructured":"Jungho Lee, Minhyeok Lee, Dogyoon Lee, and Sangyoon Lee. 2022. Hierarchically Decomposed Graph Convolutional Networks for Skeleton-Based Action Recognition. (Aug 2022)."},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_9_1","DOI":"10.1109\/lsp.2017.2678539"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_10_1","DOI":"10.1109\/ICCV48922.2021.01315"},{"unstructured":"Yanqiu Li Yanan Liu Hao Zhang Shouzheng Sun and Dan Xu. [n. d.]. Skeleton-based Human Action Recognition via Multi-Knowledge Flow Embedding Hierarchically Decomposed Graph Convolutional Network. ([n. d.]).","key":"e_1_3_2_1_11_1"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_12_1","DOI":"10.1109\/tpami.2019.2916873"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_13_1","DOI":"10.1109\/TVCG.2023.3247075"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_14_1","DOI":"10.1016\/j.knosys.2022.108146"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_15_1","DOI":"10.1109\/ICCV.2019.00554"},{"key":"e_1_3_2_1_16_1","volume-title":"Computer Vision \u2013 ECCV","author":"Ning Ma","year":"2022","unstructured":"Ma Ning, Zhang Hongyi, Li Xuhui, Zhou Sheng, Zhang Zhen, Wen Jun, Gu Jingjun, Li Haifeng, and Bu Jiajun. 2022. Learning Spatial-Preserved Skeleton Representations for Few-Shot Action Recognition. In Computer Vision \u2013 ECCV 2022. Springer International Publishing."},{"unstructured":"Helei Qiu Biao Hou Bo Ren and Xiaohua Zhang. [n. d.]. Spatio-Temporal Tuples Transformer for Skeleton-Based Action Recognition. ([n. d.]).","key":"e_1_3_2_1_17_1"},{"key":"e_1_3_2_1_18_1","article-title":"Histogram of Oriented Principal Components for Cross-View Action Recognition","author":"Rahmani Hossein","year":"2014","unstructured":"Hossein Rahmani, Arif Mahmood, DuQ. Huynh, and Ajmal Mian. 2014. Histogram of Oriented Principal Components for Cross-View Action Recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence,IEEE Transactions on Pattern Analysis and Machine Intelligence (Sep 2014).","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence,IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"e_1_3_2_1_19_1","volume-title":"HOPC: Histogram of Oriented Principal Components of 3D Pointclouds for Action Recognition","author":"Rahmani Hossein","year":"2014","unstructured":"Hossein Rahmani, Arif Mahmood, DuQ. Huynh, and Ajmal Mian. 2014. HOPC: Histogram of Oriented Principal Components of 3D Pointclouds for Action Recognition. Cornell University - arXiv,Cornell University - arXiv (Aug 2014)."},{"key":"e_1_3_2_1_20_1","volume-title":"A Large Scale Dataset for 3D Human Activity Analysis","author":"Shahroudy Amir","year":"2016","unstructured":"Amir Shahroudy, Jun Li, Tian-Tsong Ng, and Gang Wang. 2016. NTU RGB+D: A Large Scale Dataset for 3D Human Activity Analysis. Cornell University - arXiv,Cornell University - arXiv (Apr 2016)."},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_21_1","DOI":"10.1109\/CVPR.2018.00790"},{"key":"e_1_3_2_1_22_1","volume-title":"Multi-Instrumentalist Net: Unsupervised Generation of Music from Body Movements","author":"Su Kun","year":"2020","unstructured":"Kun Su, Xiulong Liu, and Eli Shlizerman. 2020. Multi-Instrumentalist Net: Unsupervised Generation of Music from Body Movements.Cornell University - arXiv,Cornell University - arXiv (Dec 2020)."},{"key":"e_1_3_2_1_23_1","volume-title":"DeepDance: Music-to-Dance Motion Choreography with Adversarial Learning. Elements,Elements (Aug","author":"Sun Guofei","year":"2020","unstructured":"Guofei Sun, Yongkang Wong, Zhiyong Cheng, MohanS. Kankanhalli, Geng Wang, and Xiangdong Li. 2020. DeepDance: Music-to-Dance Motion Choreography with Adversarial Learning. Elements,Elements (Aug 2020)."},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_24_1","DOI":"10.1145\/3240508.3240526"},{"unstructured":"Shuhei Tsuchida Satoru Fukayama Masahiro Hamasaki and Masataka Goto. 2019. AIST Dance Video Database: Multi-Genre Multi-Dancer and Multi-Camera Database for Dance Information Processing.International Symposium\/Conference on Music Information Retrieval International Symposium\/Conference on Music Information Retrieval (Jan 2019).","key":"e_1_3_2_1_25_1"},{"key":"e_1_3_2_1_26_1","volume-title":"Attention is All you Need. Neural Information Processing Systems,Neural Information Processing Systems (Jun","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, AidanN. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is All you Need. Neural Information Processing Systems,Neural Information Processing Systems (Jun 2017)."},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_27_1","DOI":"10.1109\/ICCV.2015.460"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_28_1","DOI":"10.1109\/CVPR.2014.339"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_29_1","DOI":"10.1145\/2647868.2654912"},{"doi-asserted-by":"crossref","unstructured":"Sijie Yan Yuanjun Xiong and Dahua Lin. 2018. Spatial Temporal Graph Convolutional Networks for Skeleton-Based Action Recognition. In AAAI.","key":"e_1_3_2_1_30_1","DOI":"10.1609\/aaai.v32i1.12328"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_31_1","DOI":"10.1145\/3394171.3414005"},{"key":"e_1_3_2_1_32_1","volume-title":"Hypergraph Transformer for Skeleton-based Action Recognition. (Nov","author":"Zhou Yuxuan","year":"2022","unstructured":"Yuxuan Zhou, Chao Li, Zhi-Qi Cheng, Yifeng Geng, Xuansong Xie, and Margret Keuper. 2022. Hypergraph Transformer for Skeleton-based Action Recognition. (Nov 2022)."},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_33_1","DOI":"10.1609\/aaai.v30i1.10451"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_34_1","DOI":"10.1145\/3485664"},{"key":"e_1_3_2_1_35_1","volume-title":"The Kinetics Human Action Video Dataset","author":"Zisserman Andrew","year":"2017","unstructured":"Andrew Zisserman, Joao Carreira, Karen Simonyan, Will Kay, Brian Zhang, Chloe Hillier, Sudheendra Vijayanarasimhan, Fabio Viola, Tim Green, Trevor Back, Paul Natsev, and Mustafa Suleyman. 2017. The Kinetics Human Action Video Dataset. Cornell University - arXiv,Cornell University - arXiv (May 2017)."}],"event":{"acronym":"CNIOT 2024","name":"CNIOT 2024: 2024 5th International Conference on Computing, Networks and Internet of Things","location":"Tokyo Japan"},"container-title":["Proceedings of the 2024 5th International Conference on Computing, Networks and Internet of Things"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3670105.3670151","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3670105.3670151","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T15:52:43Z","timestamp":1755877963000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3670105.3670151"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,24]]},"references-count":35,"alternative-id":["10.1145\/3670105.3670151","10.1145\/3670105"],"URL":"https:\/\/doi.org\/10.1145\/3670105.3670151","relation":{},"subject":[],"published":{"date-parts":[[2024,5,24]]},"assertion":[{"value":"2024-07-29","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}