{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T21:04:14Z","timestamp":1776891854967,"version":"3.51.2"},"reference-count":206,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100004826","name":"Beijing Natural Science Foundation","doi-asserted-by":"publisher","award":["L233026"],"award-info":[{"award-number":["L233026"]}],"id":[{"id":"10.13039\/501100004826","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62277001"],"award-info":[{"award-number":["62277001"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U25A20446"],"award-info":[{"award-number":["U25A20446"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Neurocomputing"],"published-print":{"date-parts":[[2026,6]]},"DOI":"10.1016\/j.neucom.2026.133466","type":"journal-article","created":{"date-parts":[[2026,3,27]],"date-time":"2026-03-27T08:14:56Z","timestamp":1774599296000},"page":"133466","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["Music-driven dance generation: A comprehensive review"],"prefix":"10.1016","volume":"682","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1503-7116","authenticated-orcid":false,"given":"Li","family":"Sun","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4861-0513","authenticated-orcid":false,"given":"Haisheng","family":"Li","sequence":"additional","affiliation":[]},{"given":"Fei","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Ying","family":"Liu","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.neucom.2026.133466_bib0005","series-title":"Proceedings of the 2014 International Workshop on Movement and Computing","first-page":"1","article-title":"Choreography as mediated through compositional tools for movement: constructing a historical perspective","author":"Alaoui","year":"2014"},{"key":"10.1016\/j.neucom.2026.133466_bib0010","first-page":"26","article-title":"Groovenet: real-time music-driven dance movement generation using artificial neural networks","volume":"8","author":"Alemi","year":"2017","journal-title":"networks"},{"key":"10.1016\/j.neucom.2026.133466_bib0015","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3592458","article-title":"Listen, Denoise, action! audio-driven motion synthesis with diffusion models","volume":"42","author":"Alexanderson","year":"2023","journal-title":"ACM Trans. Graph."},{"key":"10.1016\/j.neucom.2026.133466_bib0020","series-title":"Proceedings of the 30th ACM International Conference on Multimedia","first-page":"3917","article-title":"Choreograph: music-conditioned automatic dance choreography over a style and tempo consistent dynamic graph","author":"Au","year":"2022"},{"key":"10.1016\/j.neucom.2026.133466_bib0025","doi-asserted-by":"crossref","first-page":"771","DOI":"10.1007\/s11633-023-1478-9","article-title":"Rechoreonet: repertoire-based dance re-choreography with music-conditioned temporal and style clues","volume":"21","author":"Au","year":"2024","journal-title":"Mach. Intell. Res."},{"key":"10.1016\/j.neucom.2026.133466_bib0030","series-title":"Proceedings of the 24th International Conference on Artificial Intelligence","first-page":"2431","article-title":"Kinetic imaginations: exploring the possibilities of combining AI and dance","author":"Berman","year":"2015"},{"key":"10.1016\/j.neucom.2026.133466_bib0035","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"783","article-title":"Danceanyway: synthesizing beat-guided 3d dances with randomized temporal contrastive learning","author":"Bhattacharya","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0040","series-title":"Proceedings of the 8th International Conference on Movement and Computing","first-page":"1","article-title":"Generative dance-a taxonomy and survey","author":"Bisig","year":"2022"},{"key":"10.1016\/j.neucom.2026.133466_bib0045","doi-asserted-by":"crossref","first-page":"21","DOI":"10.1080\/14626260902867915","article-title":"What is generative art?","volume":"20","author":"Boden","year":"2009","journal-title":"Digit. Creat."},{"key":"10.1016\/j.neucom.2026.133466_bib0050","doi-asserted-by":"crossref","DOI":"10.1016\/j.newideapsych.2021.100902","article-title":"Group dancing as the evolutionary origin of rhythmic entrainment in humans","volume":"64","author":"Brown","year":"2022","journal-title":"New Ideas Psychol."},{"key":"10.1016\/j.neucom.2026.133466_bib0055","doi-asserted-by":"crossref","first-page":"78","DOI":"10.1038\/scientificamerican0708-78","article-title":"The Neuroscience of dance","volume":"299","author":"Brown","year":"2008","journal-title":"Sci. Am."},{"key":"10.1016\/j.neucom.2026.133466_bib0060","series-title":"International Symposium on Information and Communication Technology","first-page":"368","article-title":"Danceduo: bridging human movement and AI choreography","author":"Bui-Le","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0065","author":"Cao"},{"key":"10.1016\/j.neucom.2026.133466_bib0070","series-title":"Proceedings 20th Eurographics UK Conference","first-page":"38","article-title":"Music-driven motion editing: local motion transformations guided by music analysis","author":"Cardle","year":"2002"},{"key":"10.1016\/j.neucom.2026.133466_bib0075","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"5933","article-title":"Everybody dance now","author":"Chan","year":"2019"},{"key":"10.1016\/j.neucom.2026.133466_bib0080","doi-asserted-by":"crossref","first-page":"199","DOI":"10.1016\/j.humov.2018.12.005","article-title":"Multi-person and multisensory synchronization during group dancing","volume":"63","author":"Chauvign\u00e9","year":"2019","journal-title":"Hum. Mov. Sci."},{"key":"10.1016\/j.neucom.2026.133466_bib0085","first-page":"1","article-title":"Choreomaster: choreography-oriented music-driven dance synthesis","volume":"40","author":"Chen","year":"2021","journal-title":"ACM Trans. Graph."},{"key":"10.1016\/j.neucom.2026.133466_bib0090","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"10602","article-title":"X-dancer: expressive music to human dance video generation","author":"Chen","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0095","author":"Chu"},{"key":"10.1016\/j.neucom.2026.133466_bib0100","doi-asserted-by":"crossref","first-page":"17","DOI":"10.1038\/s43586-024-00294-7","article-title":"Graph neural networks","volume":"4","author":"Corso","year":"2024","journal-title":"Nat. Rev. Methods Primers"},{"key":"10.1016\/j.neucom.2026.133466_bib0105","author":"Crnkovic-Friis"},{"key":"10.1016\/j.neucom.2026.133466_bib0110","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"9760","article-title":"Mofusion: a framework for denoising-diffusion-based motion synthesis","author":"Dabral","year":"2023"},{"key":"10.1016\/j.neucom.2026.133466_bib0115","doi-asserted-by":"crossref","first-page":"61","DOI":"10.1007\/s11263-025-02611-3","article-title":"Tcdiff++: an end-to-end trajectory-controllable diffusion model for harmonious music-driven group choreography","volume":"134","author":"Dai","year":"2026","journal-title":"Int. J. Comput. Vis."},{"key":"10.1016\/j.neucom.2026.133466_bib0120","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"2645","article-title":"Harmonious music-driven group choreography with trajectory-controllable diffusion","author":"Dai","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0125","doi-asserted-by":"crossref","DOI":"10.1016\/j.heliyon.2022.e12750","article-title":"The computer, a choreographer? Aesthetic responses to randomly-generated dance choreography by a computer","volume":"9","author":"Darda","year":"2023","journal-title":"Heliyon"},{"key":"10.1016\/j.neucom.2026.133466_bib0130","series-title":"Proc. Interspeech 2025","first-page":"186","article-title":"Ffd: fine-finger diffusion model for music to fine-grained finger dance generation","author":"Dong","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0135","author":"Dong"},{"key":"10.1016\/j.neucom.2026.133466_bib0140","doi-asserted-by":"crossref","DOI":"10.1371\/journal.pone.0164783","article-title":"One in the dance: musical correlates of group synchrony in a real-world club environment","volume":"11","author":"Ellamil","year":"2016","journal-title":"PLoS One"},{"key":"10.1016\/j.neucom.2026.133466_bib0145","first-page":"157","article-title":"A multi-modal dance corpus for research into interaction between humans in virtual environments","volume":"7","author":"Essid","year":"2013","journal-title":"J. Multimodal User Interfaces"},{"key":"10.1016\/j.neucom.2026.133466_bib0150","doi-asserted-by":"crossref","DOI":"10.1016\/j.compeleceng.2022.108310","article-title":"A bi-directional attention guided cross-modal network for music based dance generation","volume":"103","author":"Fan","year":"2022","journal-title":"Comput. Electr. Eng."},{"key":"10.1016\/j.neucom.2026.133466_bib0155","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"13336","article-title":"Go to zero: towards zero-shot motion generation with million-scale data","author":"Fan","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0160","first-page":"501","article-title":"Example-based automatic music-driven conventional dance motion synthesis","volume":"18","author":"Fan","year":"2011","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"10.1016\/j.neucom.2026.133466_bib0165","author":"Fan"},{"key":"10.1016\/j.neucom.2026.133466_bib0170","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"1688","article-title":"Everything2motion: synchronizing diverse inputs via a unified framework for human motion synthesis","author":"Fan","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0175","doi-asserted-by":"crossref","first-page":"11","DOI":"10.1016\/j.cag.2020.09.009","article-title":"Learning to dance: a graph convolutional adversarial network to generate realistic dance motions from audio","volume":"94","author":"Ferreira","year":"2021","journal-title":"Comput. Graph."},{"key":"10.1016\/j.neucom.2026.133466_bib0180","doi-asserted-by":"crossref","first-page":"351","DOI":"10.1016\/j.evolhumbehav.2021.01.003","article-title":"Evolution and functions of human dance","volume":"42","author":"Fink","year":"2021","journal-title":"Evolution and Human Behavior"},{"key":"10.1016\/j.neucom.2026.133466_bib0185","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3728633","article-title":"Ai-generated content (AIGC) for various data modalities: a survey","volume":"57","author":"Foo","year":"2025","journal-title":"ACM Comput. Surv."},{"key":"10.1016\/j.neucom.2026.133466_bib0190","author":"Fu"},{"key":"10.1016\/j.neucom.2026.133466_bib0195","first-page":"177","article-title":"Music content driven automated choreography with beat-wise motion connectivity constraints","author":"Fukayama","year":"2015","journal-title":"Proc. SMC"},{"key":"10.1016\/j.neucom.2026.133466_bib0200","series-title":"Understanding Generative AI in a Cultural Context: Artificial Myths and Human Realities","first-page":"51","article-title":"Algorithmic choreography: the kinetic fusion of dance and generative AI","author":"Furiasse","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0205","doi-asserted-by":"crossref","first-page":"146","DOI":"10.1002\/9781118475249.ch5","article-title":"Generative art theory","author":"Galanter","year":"2016","journal-title":"A companion to digital art"},{"key":"10.1016\/j.neucom.2026.133466_bib0210","series-title":"Proceedings of EVA London 2019","article-title":"Artificial intelligence and problems in generative art theory","author":"Galanter","year":"2019"},{"key":"10.1016\/j.neucom.2026.133466_bib0215","series-title":"Proceedings of the 30th ACM International Conference on Multimedia","first-page":"1261","article-title":"Pc-dance: posture-controllable music-driven dance synthesis","author":"Gao","year":"2022"},{"key":"10.1016\/j.neucom.2026.133466_bib0220","author":"Gao"},{"key":"10.1016\/j.neucom.2026.133466_bib0225","series-title":"Proceedings of the Special Interest Group on Computer Graphics and Interactive Techniques Conference Conference Papers","first-page":"1","article-title":"Duetgen: music driven two-person dance generation via hierarchical masked modeling","author":"Ghosh","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0230","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"9942","article-title":"Tm2d: bimodality driven 3d dance generation via music-text integration","author":"Gong","year":"2023"},{"key":"10.1016\/j.neucom.2026.133466_bib0235","series-title":"Choreographics: A Comparison of Dance Notation Systems from the Fifteenth Century to the Present","author":"Guest","year":"2014"},{"key":"10.1016\/j.neucom.2026.133466_bib0240","doi-asserted-by":"crossref","first-page":"5559","DOI":"10.1109\/TIP.2021.3086082","article-title":"Danceit: music-inspired dancing video synthesis","volume":"30","author":"Guo","year":"2021","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.neucom.2026.133466_bib0245","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"13932","article-title":"Mdd: a dataset for text-and-music conditioned duet dance generation","author":"Gupta","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0250","author":"Han"},{"key":"10.1016\/j.neucom.2026.133466_bib0255","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3414685.3417836","article-title":"Moglow: probabilistic and controllable motion synthesis using normalising flows","volume":"39","author":"Henter","year":"2020","journal-title":"ACM Trans. Graph."},{"key":"10.1016\/j.neucom.2026.133466_bib0260","article-title":"Gans trained by a two time-scale update rule converge to a local nash equilibrium","volume":"30","author":"Heusel","year":"2017","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.neucom.2026.133466_bib0265","first-page":"6840","article-title":"Denoising diffusion probabilistic models","volume":"33","author":"Ho","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.neucom.2026.133466_bib0270","series-title":"2023 IEEE International Conference on Multimedia and Expo (ICME)","first-page":"2111","article-title":"Graph convolutional GRU for music-oriented dance choreography generation","author":"Hou","year":"2023"},{"key":"10.1016\/j.neucom.2026.133466_bib0275","series-title":"International Conference on Multimedia Modeling","first-page":"243","article-title":"Tg-Dance: transgan-based intelligent Dance generation with music","author":"Huang","year":"2023"},{"key":"10.1016\/j.neucom.2026.133466_bib0280","series-title":"International Conference on Learning Representations","article-title":"Dance Revolution: long-term dance generation with music via curriculum learning","author":"Huang","year":"2021"},{"key":"10.1016\/j.neucom.2026.133466_bib0285","doi-asserted-by":"crossref","first-page":"9817","DOI":"10.1007\/s00521-021-05752-x","article-title":"Choreography CGAN: generating dances with music beats using conditional generative adversarial networks","volume":"33","author":"Huang","year":"2021","journal-title":"Neural Comput. Appl."},{"key":"10.1016\/j.neucom.2026.133466_bib0290","series-title":"European Conference on Computer Vision","first-page":"273","article-title":"Beat-it: beat-synchronized multi-condition 3d dance generation","author":"Huang","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0295","first-page":"2","article-title":"Sdcgm: dance choreography generation from music with dancer-specific style","volume":"34","author":"Ishii","year":"2026","journal-title":"J. Inf. Process."},{"key":"10.1016\/j.neucom.2026.133466_bib0300","first-page":"20067","article-title":"Motiongpt: human motion as a foreign language","volume":"36","author":"Jiang","year":"2023","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.neucom.2026.133466_bib0305","first-page":"1073","article-title":"Sensor based dance coherent action generation model using deep learning framework","volume":"25","author":"Jiang","year":"2024","journal-title":"Scalable Comput.: Pract. Exp."},{"key":"10.1016\/j.neucom.2026.133466_bib0310","doi-asserted-by":"crossref","DOI":"10.1098\/rspa.2021.0071","article-title":"An extensive review of computational dance Automation techniques and applications","volume":"477","author":"Joshi","year":"2021","journal-title":"Proc. R. Soc. A"},{"key":"10.1016\/j.neucom.2026.133466_bib0315","series-title":"2023 IEEE International Conference on Multimedia and Expo (ICME)","first-page":"31","article-title":"Action-gpt: leveraging large-scale language models for improved and generalized action generation","author":"Kalakonda","year":"2023"},{"key":"10.1016\/j.neucom.2026.133466_bib0320","doi-asserted-by":"crossref","first-page":"15036","DOI":"10.1109\/TPAMI.2023.3312092","article-title":"Mnet++: music-driven pluralistic dancing toward multiple dance genre synthesis","volume":"45","author":"Kim","year":"2023","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.neucom.2026.133466_bib0325","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"3490","article-title":"A brand new dance partner: music-conditioned pluralistic dancing controlled by multiple dance genres","author":"Kim","year":"2022"},{"key":"10.1016\/j.neucom.2026.133466_bib0330","series-title":"AAAI Fall Symposium: Aurally Informed Performance","first-page":"2","article-title":"Making them dance","author":"Kim","year":"2006"},{"key":"10.1016\/j.neucom.2026.133466_bib0335","doi-asserted-by":"crossref","first-page":"375","DOI":"10.1002\/cav.314","article-title":"Perceptually motivated automatic dance motion generation for music","volume":"20","author":"Kim","year":"2009","journal-title":"Comput. Animat. Virtual Worlds"},{"key":"10.1016\/j.neucom.2026.133466_bib0340","doi-asserted-by":"crossref","first-page":"392","DOI":"10.1145\/882262.882283","article-title":"Rhythmic-motion synthesis based on motion-beat analysis","volume":"22","author":"Kim","year":"2003","journal-title":"ACM Trans. Graph."},{"key":"10.1016\/j.neucom.2026.133466_bib0345","author":"Kodaira"},{"key":"10.1016\/j.neucom.2026.133466_bib0350","doi-asserted-by":"crossref","first-page":"44982","DOI":"10.1109\/ACCESS.2022.3169782","article-title":"Danceconv: Dance motion generation with convolutional networks","volume":"10","author":"Kritsis","year":"2022","journal-title":"Ieee Access"},{"key":"10.1016\/j.neucom.2026.133466_bib0355","series-title":"Proceedings of the 7th Annual Workshop on Genetic and Evolutionary Computation","first-page":"366","article-title":"Choreogenetics: the generation of choreographic variants through genetic mutations and selection","author":"Lapointe","year":"2005"},{"key":"10.1016\/j.neucom.2026.133466_bib0360","series-title":"European Conference on Computer Vision","first-page":"293","article-title":"Scalable group choreography via variational phase manifold learning","author":"Le","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0365","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3618356","article-title":"Controllable group choreography using contrastive diffusion","volume":"42","author":"Le","year":"2023","journal-title":"ACM Trans. on Graph. (TOG)"},{"key":"10.1016\/j.neucom.2026.133466_bib0370","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"8673","article-title":"Music-driven group choreography","author":"Le","year":"2023"},{"key":"10.1016\/j.neucom.2026.133466_bib0375","doi-asserted-by":"crossref","first-page":"436","DOI":"10.1038\/nature14539","article-title":"Deep learning","volume":"521","author":"LeCun","year":"2015","journal-title":"nature"},{"key":"10.1016\/j.neucom.2026.133466_bib0380","series-title":"Computer Graphics Forum","first-page":"353","article-title":"Automatic synchronization of background music and motion in computer animation","author":"Lee","year":"2005"},{"key":"10.1016\/j.neucom.2026.133466_bib0385","article-title":"Dancing to music","volume":"32","author":"Lee","year":"2019","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.neucom.2026.133466_bib0390","author":"Lee"},{"key":"10.1016\/j.neucom.2026.133466_bib0395","series-title":"20th International Society for Music Information Retrieval Conference, ISMIR 2019","first-page":"894","article-title":"Automatic choreography generation with convolutional encoder-decoder network","author":"Lee","year":"2019"},{"key":"10.1016\/j.neucom.2026.133466_bib0400","doi-asserted-by":"crossref","first-page":"139","DOI":"10.1007\/s11370-017-0241-x","article-title":"Dance motion generation by recombination of body parts from motion source","volume":"11","author":"Lee","year":"2018","journal-title":"Intell. Serv. Robot."},{"key":"10.1016\/j.neucom.2026.133466_bib0405","doi-asserted-by":"crossref","first-page":"895","DOI":"10.1007\/s11042-012-1288-5","article-title":"Music similarity-based approach to generating dance motion sequence","volume":"62","author":"Lee","year":"2013","journal-title":"Multimed. Tools Appl."},{"key":"10.1016\/j.neucom.2026.133466_bib0410","author":"Li"},{"key":"10.1016\/j.neucom.2026.133466_bib0415","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"1272","article-title":"Danceformer: music conditioned 3d dance generation with parametric motion transformer","author":"Li","year":"2022"},{"key":"10.1016\/j.neucom.2026.133466_bib0420","author":"Li"},{"key":"10.1016\/j.neucom.2026.133466_bib0425","doi-asserted-by":"crossref","first-page":"1488","DOI":"10.1109\/TMM.2021.3066115","article-title":"Rhythm-aware sequence-to-sequence learning for labanotation generation with gesture-sensitive graph convolutional encoding","volume":"24","author":"Li","year":"2021","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.neucom.2026.133466_bib0430","series-title":"ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","first-page":"8281","article-title":"Exploring multi-modal control in music-driven dance generation","author":"Li","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0435","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"13401","article-title":"AI choreographer: music conditioned 3d dance generation with aist++","author":"Li","year":"2021"},{"key":"10.1016\/j.neucom.2026.133466_bib0440","first-page":"1","article-title":"Lodge++: high-quality and long dance generation with vivid choreography patterns","author":"Li","year":"2025","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.neucom.2026.133466_bib0445","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"1524","article-title":"Lodge: a coarse to fine diffusion network for long dance generation guided by the characteristic dance primitives","author":"Li","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0450","author":"Li"},{"key":"10.1016\/j.neucom.2026.133466_bib0455","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"10234","article-title":"Finedance: a fine-grained choreography dataset for 3d full body dance generation","author":"Li","year":"2023"},{"key":"10.1016\/j.neucom.2026.133466_bib0460","author":"Li"},{"key":"10.1016\/j.neucom.2026.133466_bib0465","author":"Li"},{"key":"10.1016\/j.neucom.2026.133466_bib0470","first-page":"1","article-title":"Example-based motion synthesis via generative motion matching","volume":"42","author":"Li","year":"2023","journal-title":"ACM Trans. on Graph. (TOG)"},{"key":"10.1016\/j.neucom.2026.133466_bib0475","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"14420","article-title":"Music-aligned holistic 3d dance generation via hierarchical motion modeling","author":"Li","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0480","series-title":"Proceedings of the Computer Vision and Pattern Recognition Conference","first-page":"27805","article-title":"Unipose: a unified multimodal framework for human pose comprehension, generation and editing","author":"Li","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0485","series-title":"Proceedings of the 2025 International Conference on Multimedia Retrieval","first-page":"671","article-title":"Melodance: dance generation guided by music structure and emotion","author":"Li","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0490","doi-asserted-by":"crossref","first-page":"3463","DOI":"10.1007\/s11263-024-02042-6","article-title":"Intergen: diffusion-based multi-human motion generation under complex interactions","volume":"132","author":"Liang","year":"2024","journal-title":"Int. J. Comput. Vis."},{"key":"10.1016\/j.neucom.2026.133466_bib0495","author":"Lin"},{"key":"10.1016\/j.neucom.2026.133466_bib0500","doi-asserted-by":"crossref","first-page":"9330","DOI":"10.1109\/TMM.2024.3390232","article-title":"Music-driven choreography based on music feature clusters and dynamic programming","volume":"26","author":"Lin","year":"2024","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.neucom.2026.133466_bib0505","author":"Ling"},{"key":"10.1016\/j.neucom.2026.133466_bib0510","author":"Ling"},{"key":"10.1016\/j.neucom.2026.133466_bib0515","author":"Liu"},{"key":"10.1016\/j.neucom.2026.133466_bib0520","author":"Liu"},{"key":"10.1016\/j.neucom.2026.133466_bib0525","author":"Liu"},{"key":"10.1016\/j.neucom.2026.133466_bib0530","series-title":"Proceedings of the 2024 ACM Designing Interactive Systems Conference","first-page":"920","article-title":"Dancegen: supporting choreography ideation and prototyping with generative AI","author":"Liu","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0535","first-page":":248:1","article-title":"Smpl: a skinned multi-person linear model","volume":"34","author":"Loper","year":"2015","journal-title":"ACM Trans. Graphics (Proc. SIGGRAPH Asia)"},{"key":"10.1016\/j.neucom.2026.133466_bib0540","series-title":"SMPL: A Skinned Multi-Person Linear Model. 1 Ed","first-page":"851","author":"Loper","year":"2023"},{"key":"10.1016\/j.neucom.2026.133466_bib0545","first-page":"28051","article-title":"M3GPT: an advanced multimodal, multitask framework for motion comprehension and generation","volume":"37","author":"Luo","year":"2024","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.neucom.2026.133466_bib0550","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"26984","article-title":"Popdg: popular 3d dance generation with popdanceset","author":"Luo","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0555","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"5442","article-title":"Amass: archive of motion capture as surface shapes","author":"Mahmood","year":"2019"},{"key":"10.1016\/j.neucom.2026.133466_bib0560","author":"Maluleke"},{"key":"10.1016\/j.neucom.2026.133466_bib0565","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"3126","article-title":"M2c: concise music representation for 3d dance generation","author":"Marchellus","year":"2023"},{"key":"10.1016\/j.neucom.2026.133466_bib0570","series-title":"2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)","first-page":"57","article-title":"Performance-driven dance motion control of a virtual partner character","author":"Mousas","year":"2018"},{"key":"10.1016\/j.neucom.2026.133466_bib0575","series-title":"IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, 2005","first-page":"275","article-title":"Chroma-based statistical audio features for audio matching","author":"Muller","year":"2005"},{"key":"10.1016\/j.neucom.2026.133466_bib0580","series-title":"The Fifth Richard Tapia Celebration of Diversity in Computing Conference: Intellect, Initiatives, Insight, and Innovations","first-page":"30","article-title":"Dancing, dance and choreography: an intelligent nondeterministic generator","author":"Nakazawa","year":"2009"},{"key":"10.1016\/j.neucom.2026.133466_bib0585","doi-asserted-by":"crossref","first-page":"60","DOI":"10.1080\/14794713.2024.2338927","article-title":"Exploring the impact of machine learning on dance performance: a systematic review","volume":"20","author":"Nogueira","year":"2024","journal-title":"Int. J. Perform. Arts Digit. Media"},{"key":"10.1016\/j.neucom.2026.133466_bib0590","series-title":"2007 IEEE International Conference on Multimedia and Expo","first-page":"1703","article-title":"Multicamera audio-visual analysis of dance figures","author":"Ofli","year":"2007"},{"key":"10.1016\/j.neucom.2026.133466_bib0595","doi-asserted-by":"crossref","first-page":"93","DOI":"10.1007\/s12193-008-0009-x","article-title":"An audio-driven dancing avatar","volume":"2","author":"Ofli","year":"2008","journal-title":"J. Multimodal User Interfaces"},{"key":"10.1016\/j.neucom.2026.133466_bib0600","series-title":"2010 IEEE International Conference on Acoustics, Speech and Signal Processing","first-page":"2466","article-title":"Multi-modal analysis of dance performances for music-driven choreography synthesis","author":"Ofli","year":"2010"},{"key":"10.1016\/j.neucom.2026.133466_bib0605","doi-asserted-by":"crossref","first-page":"747","DOI":"10.1109\/TMM.2011.2181492","article-title":"Learn2dance: learning statistical music-to-dance mappings for choreography synthesis","volume":"14","author":"Ofli","year":"2011","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.neucom.2026.133466_bib0610","article-title":"Fmdistance: a fast and effective distance function for motion capture data","volume":"7","author":"Onuma","year":"2008","journal-title":"Eurographics (Short Papers)"},{"key":"10.1016\/j.neucom.2026.133466_bib0615","author":"Pang"},{"key":"10.1016\/j.neucom.2026.133466_bib0620","doi-asserted-by":"crossref","DOI":"10.1080\/08839514.2023.2226962","article-title":"Dance video motion recognition based on computer vision and image processing","volume":"37","author":"Pang","year":"2023","journal-title":"Appl. Artif. Intell."},{"key":"10.1016\/j.neucom.2026.133466_bib0625","first-page":"1","article-title":"Normalizing flows for probabilistic modeling and inference","volume":"22","author":"Papamakarios","year":"2021","journal-title":"J. Mach. Learn. Res."},{"key":"10.1016\/j.neucom.2026.133466_bib0630","series-title":"International Conference on ArtsIT, Interactivity and Game Creation","first-page":"447","article-title":"Pirounet: creating dance through artist-centric deep learning","author":"Papillon","year":"2022"},{"key":"10.1016\/j.neucom.2026.133466_bib0635","series-title":"Proceedings of the 31st ACM International Conference on Multimedia","first-page":"1374","article-title":"Diffdance: cascaded human motion diffusion model for dance generation","author":"Qi","year":"2023"},{"key":"10.1016\/j.neucom.2026.133466_bib0640","doi-asserted-by":"crossref","first-page":"166540","DOI":"10.1109\/ACCESS.2019.2953698","article-title":"Music-driven dance generation","volume":"7","author":"Qi","year":"2019","journal-title":"Ieee Access"},{"key":"10.1016\/j.neucom.2026.133466_bib0645","author":"Qiang"},{"key":"10.1016\/j.neucom.2026.133466_bib0650","series-title":"Proceedings of the 2011 ACM SIGGRAPH\/Eurographics Symposium on Computer Animation","first-page":"147","article-title":"Real-time classification of dance gestures from skeleton animation","author":"Raptis","year":"2011"},{"key":"10.1016\/j.neucom.2026.133466_bib0655","series-title":"Proceedings of the 28th ACM International Conference on Multimedia","first-page":"46","article-title":"Self-supervised dance video synthesis conditioned on music","author":"Ren","year":"2020"},{"key":"10.1016\/j.neucom.2026.133466_bib0660","article-title":"Cultural heritage preservation through dance digitization: a review","volume":"28","author":"Reshma","year":"2023","journal-title":"Digit. Appl. Archaeol. Cult. Herit."},{"key":"10.1016\/j.neucom.2026.133466_bib0665","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1080\/01472526.2019.1575661","article-title":"Information technology and the arts: the evolution of computer choreography during the last half century","volume":"42","author":"Sagasti","year":"2019","journal-title":"Dance Chron."},{"key":"10.1016\/j.neucom.2026.133466_bib0670","author":"Shah"},{"key":"10.1016\/j.neucom.2026.133466_bib0675","article-title":"Understanding dance semantics using spatio-temporal features coupled GRU networks","volume":"42","author":"Shailesh","year":"2022","journal-title":"Entertain. Comput."},{"key":"10.1016\/j.neucom.2026.133466_bib0680","series-title":"Computer Graphics Forum","first-page":"449","article-title":"Dancing-to-music character animation","author":"Shiratori","year":"2006"},{"key":"10.1016\/j.neucom.2026.133466_bib0685","series-title":"Proceedings 2006 IEEE International Conference on Robotics and Automation, 2006. ICRA 2006","first-page":"3654","article-title":"Synthesizing dance performance using musical and motion features","author":"Shiratori","year":"2006"},{"key":"10.1016\/j.neucom.2026.133466_bib0690","author":"Siyao"},{"key":"10.1016\/j.neucom.2026.133466_bib0695","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"11050","article-title":"Bailando: 3d dance generation by actor-critic GPT with choreographic memory","author":"Siyao","year":"2022"},{"key":"10.1016\/j.neucom.2026.133466_bib0700","doi-asserted-by":"crossref","first-page":"14192","DOI":"10.1109\/TPAMI.2023.3319435","article-title":"Bailando++: 3d dance GPT with choreographic memory","volume":"45","author":"Siyao","year":"2023","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.neucom.2026.133466_bib0705","doi-asserted-by":"crossref","first-page":"111","DOI":"10.1177\/1029864917712345","article-title":"Group behaviour and interpersonal synchronization to electronic dance music","volume":"23","author":"Solberg","year":"2019","journal-title":"Musicae Scientiae"},{"key":"10.1016\/j.neucom.2026.133466_bib0710","series-title":"Euro-Mediterranean Conference","first-page":"404","article-title":"Digitization of Cypriot folk dances","author":"Stavrakis","year":"2012"},{"key":"10.1016\/j.neucom.2026.133466_bib0715","author":"Sui"},{"key":"10.1016\/j.neucom.2026.133466_bib0720","doi-asserted-by":"crossref","first-page":"497","DOI":"10.1109\/TMM.2020.2981989","article-title":"Deepdance: music-to-dance motion choreography with adversarial learning","volume":"23","author":"Sun","year":"2020","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.neucom.2026.133466_bib0725","series-title":"Proceedings of the 26th ACM International Conference on Multimedia","first-page":"1598","article-title":"Dance with melody: an lstm-autoencoder approach to music-oriented dance synthesis","author":"Tang","year":"2018"},{"key":"10.1016\/j.neucom.2026.133466_bib0730","series-title":"Proceedings of the 26th ACM International Conference on Multimedia","first-page":"1237","article-title":"Anidance: real-time dance motion synthesize to the song","author":"Tang","year":"2018"},{"key":"10.1016\/j.neucom.2026.133466_bib0735","first-page":"16083","article-title":"Any-to-any generation via composable diffusion","volume":"36","author":"Tang","year":"2023","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.neucom.2026.133466_bib0740","series-title":"Proceedings of the Asian Conference on Computer Vision","first-page":"2667","article-title":"Rethinking sampling for music-driven long-term dance generation","author":"Truong-Thuy","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0745","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"448","article-title":"Edge: editable dance generation from music","author":"Tseng","year":"2023"},{"key":"10.1016\/j.neucom.2026.133466_bib0750","series-title":"International Conference on Multimedia Modeling","first-page":"251","article-title":"Query-by-dancing: a dance music retrieval system based on body-motion similarity","author":"Tsuchida","year":"2018"},{"key":"10.1016\/j.neucom.2026.133466_bib0755","series-title":"ISMIR","first-page":"6","article-title":"Aist dance video database: multi-genre, multi-dancer, and multi-camera database for dance information processing","author":"Tsuchida","year":"2019"},{"key":"10.1016\/j.neucom.2026.133466_bib0760","doi-asserted-by":"crossref","first-page":"59","DOI":"10.2307\/1575623","article-title":"Nuntius: a computer system for the interactive composition and analysis of music and dance","volume":"25","author":"Ungvary","year":"1992","journal-title":"Leonardo"},{"key":"10.1016\/j.neucom.2026.133466_bib0765","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3478513.3480570","article-title":"Transflower: probabilistic autoregressive dance generation with multimodal attention","volume":"40","author":"Valle-P\u00e9rez","year":"2021","journal-title":"ACM Trans. Graph."},{"key":"10.1016\/j.neucom.2026.133466_bib0770","series-title":"11th International Conference on Computational Creativity","first-page":"284","article-title":"Towards movement generation with audio features","author":"Wallace","year":"2020"},{"key":"10.1016\/j.neucom.2026.133466_bib0775","doi-asserted-by":"crossref","first-page":"113","DOI":"10.1038\/s40494-025-01668-0","article-title":"Discrete diffusion model with contrastive learning for music to natural and long dance generation","volume":"13","author":"Wang","year":"2025","journal-title":"npj Heritage Science"},{"key":"10.1016\/j.neucom.2026.133466_bib0780","doi-asserted-by":"crossref","first-page":"588","DOI":"10.3390\/s24020588","article-title":"A music-driven dance generation method based on a spatial-temporal refinement model to optimize abnormal frames","volume":"24","author":"Wang","year":"2024","journal-title":"Sensors"},{"key":"10.1016\/j.neucom.2026.133466_bib0785","series-title":"Australasian Joint Conference on Artificial Intelligence","first-page":"277","article-title":"Midget: music conditioned 3d dance generation","author":"Wang","year":"2023"},{"key":"10.1016\/j.neucom.2026.133466_bib0790","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"8798","article-title":"High-resolution image synthesis and semantic manipulation with conditional gans","author":"Wang","year":"2018"},{"key":"10.1016\/j.neucom.2026.133466_bib0795","series-title":"Proceedings of the 33rd ACM International Conference on Multimedia","first-page":"7912","article-title":"Choreomuse: robust music-to-dance video generation with style transfer and beat-adherent motion","author":"Wang","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0800","series-title":"2025 IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV)","first-page":"5136","article-title":"Dance any beat: blending beats with visuals in dance video generation","author":"Wang","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0805","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"7892","article-title":"Dancecamera3d: 3d camera movement synthesis with music and dance","author":"Wang","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0810","series-title":"Proceedings of the 30th ACM International Conference on Multimedia","first-page":"1138","article-title":"Groupdancer: music to multi-people dance synthesis with style collaboration","author":"Wang","year":"2022"},{"key":"10.1016\/j.neucom.2026.133466_bib0815","series-title":"Proceedings of the 32nd ACM International Conference on Multimedia","first-page":"10200","article-title":"Dancecamanimator: keyframe-based controllable 3d dance camera synthesis","author":"Wang","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0820","author":"Wang"},{"key":"10.1016\/j.neucom.2026.133466_bib0825","series-title":"Proceedings of the Computer Vision and Pattern Recognition Conference","first-page":"27849","article-title":"Mg-motionllm: a unified framework for motion comprehension and generation across multiple granularities","author":"Wu","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0830","series-title":"Forty-First International Conference on Machine Learning","article-title":"Next-Gpt: any-to-any multimodal LLM","author":"Wu","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0835","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2024.109610","article-title":"Transformer-based partner dance motion generation","volume":"139","author":"Wu","year":"2025","journal-title":"Eng. Appl. Artif. Intell."},{"key":"10.1016\/j.neucom.2026.133466_bib0840","series-title":"2024 International Joint Conference on Neural Networks (IJCNN)","first-page":"1","article-title":"Music-driven character dance video generation based on pre-trained diffusion model","author":"Xu","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0845","series-title":"Proceedings of the 9th International Conference on Movement and Computing","first-page":"1","article-title":"Dancecraft: a music-reactive real-time dance improv system","author":"Xu","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0850","series-title":"2019 International Joint Conference on Neural Networks (IJCNN)","first-page":"1","article-title":"Weakly-supervised deep recurrent neural networks for basic dance step generation","author":"Yalta","year":"2019"},{"key":"10.1016\/j.neucom.2026.133466_bib0855","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","article-title":"Spatial temporal graph convolutional networks for skeleton-based action recognition","author":"Yan","year":"2018"},{"key":"10.1016\/j.neucom.2026.133466_bib0860","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"25615","article-title":"Unimumo: unified text, music, and motion generation","author":"Yang","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0865","series-title":"Proceedings of the 2024 International Conference on Multimedia Retrieval","first-page":"675","article-title":"Codancers: music-driven coherent group dance generation with choreographic unit","author":"Yang","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0870","series-title":"Proceedings of the 33rd ACM International Conference on Multimedia","first-page":"6663","article-title":"Cohedancers: enhancing interactive group dance generation through music-driven coherence decomposition","author":"Yang","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0875","series-title":"Proceedings of the 2024 International Conference on Multimedia Retrieval","first-page":"11","article-title":"Beatdance: a beat-based model-agnostic contrastive learning framework for music-dance retrieval","author":"Yang","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0880","author":"Yang"},{"key":"10.1016\/j.neucom.2026.133466_bib0885","doi-asserted-by":"crossref","first-page":"3474","DOI":"10.1109\/TVCG.2023.3235538","article-title":"Keyframe control of music-driven 3d dance generation","volume":"30","author":"Yang","year":"2023","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"10.1016\/j.neucom.2026.133466_bib0890","series-title":"Proceedings of the 31st ACM International Conference on Multimedia","first-page":"8504","article-title":"Dance with you: the diversity controllable dancer generation via diffusion models","author":"Yao","year":"2023"},{"key":"10.1016\/j.neucom.2026.133466_bib0895","doi-asserted-by":"crossref","first-page":"35","DOI":"10.1016\/j.aiopen.2021.12.002","article-title":"Human motion modeling with deep learning: a survey","volume":"3","author":"Ye","year":"2022","journal-title":"AI Open"},{"key":"10.1016\/j.neucom.2026.133466_bib0900","series-title":"Proceedings of the 28th ACM International Conference on Multimedia","first-page":"744","article-title":"Choreonet: towards music to dance synthesis with choreographic action unit","author":"Ye","year":"2020"},{"key":"10.1016\/j.neucom.2026.133466_bib0905","author":"Yin"},{"key":"10.1016\/j.neucom.2026.133466_bib0910","series-title":"Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","first-page":"9637","article-title":"Anygpt: unified multimodal LLM with discrete sequence modeling","author":"Zhan","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0915","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"687","article-title":"Bidirectional autoregessive diffusion model for dance generation","author":"Zhang","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0920","first-page":"1","article-title":"Mqenet: a mesh quality evaluation neural network based on dynamic graph attention","author":"Zhang","year":"2026","journal-title":"Eng. Comput."},{"key":"10.1016\/j.neucom.2026.133466_bib0925","doi-asserted-by":"crossref","first-page":"321","DOI":"10.1007\/s00366-024-02006-x","article-title":"Mtgnet: multi-label mesh quality evaluation using topology-guided graph neural network","volume":"41","author":"Zhang","year":"2025","journal-title":"Eng. with Comput."},{"key":"10.1016\/j.neucom.2026.133466_bib0930","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV)","first-page":"12158","article-title":"Danceeditor: towards iterative editable music-driven dance generation with open-vocabulary descriptions","author":"Zhang","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0935","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2025.126959","article-title":"Meshkinn: a self-supervised Mesh generation model based on kolmogorov\u2013arnold-informed neural network","volume":"275","author":"Zhang","year":"2025","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.neucom.2026.133466_bib0940","series-title":"Proceedings of the 33rd ACM International Conference on Multimedia","first-page":"10447","article-title":"Edmg: towards efficient long dance motion generation with fundamental movements from dance genres","author":"Zhang","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0945","doi-asserted-by":"crossref","first-page":"6023","DOI":"10.1109\/TCE.2024.3386657","article-title":"Quality guided metric learning for domain adaptation person re-identification","volume":"70","author":"Zhang","year":"2024","journal-title":"IEEE Trans. Consum. Electron."},{"key":"10.1016\/j.neucom.2026.133466_bib0950","series-title":"European Conference on Computer Vision","first-page":"397","article-title":"Large motion model for unified multi-modal motion generation","author":"Zhang","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0955","series-title":"Proceedings of the 2022 International Conference on Multimedia Retrieval","first-page":"34","article-title":"Music-to-dance generation with multiple conformer","author":"Zhang","year":"2022"},{"key":"10.1016\/j.neucom.2026.133466_bib0960","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"7368","article-title":"Motiongpt: finetuned llms are general-purpose motion generators","author":"Zhang","year":"2024"},{"key":"10.1016\/j.neucom.2026.133466_bib0965","author":"Zhang"},{"key":"10.1016\/j.neucom.2026.133466_bib0970","first-page":"1","article-title":"Deep multimodal data fusion","volume":"56","author":"Zhao","year":"2024","journal-title":"ACM Comput. Surv."},{"key":"10.1016\/j.neucom.2026.133466_bib0975","author":"Zhao"},{"key":"10.1016\/j.neucom.2026.133466_bib0980","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"10560","article-title":"Freedance: towards harmonic free-number group dance generation via a unified framework","author":"Zhao","year":"2025"},{"key":"10.1016\/j.neucom.2026.133466_bib0985","article-title":"A survey of deep learning in sports applications: perception, comprehension, and decision","author":"Zhao","year":"2025","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"10.1016\/j.neucom.2026.133466_bib0990","doi-asserted-by":"crossref","DOI":"10.1016\/j.jvcir.2022.103486","article-title":"Colorful 3d reconstruction at high resolution using multi-view representation","volume":"85","author":"Zheng","year":"2022","journal-title":"J. Vis. Commun. Image Represent."},{"key":"10.1016\/j.neucom.2026.133466_bib0995","series-title":"Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems","first-page":"1","article-title":"Dance and choreography in HCI: a two-decade retrospective","author":"Zhou","year":"2021"},{"key":"10.1016\/j.neucom.2026.133466_bib1000","doi-asserted-by":"crossref","first-page":"351","DOI":"10.1007\/s11633-021-1293-0","article-title":"Deep audio-visual learning: a survey","volume":"18","author":"Zhu","year":"2021","journal-title":"Int. J. Autom. Comput."},{"key":"10.1016\/j.neucom.2026.133466_bib1005","doi-asserted-by":"crossref","first-page":"2430","DOI":"10.1109\/TPAMI.2023.3330935","article-title":"Human motion generation: a survey","volume":"46","author":"Zhu","year":"2023","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.neucom.2026.133466_bib1010","series-title":"European Conference on Computer Vision","first-page":"182","article-title":"Quantized GAN for complex music generation from dance videos","author":"Zhu","year":"2022"},{"key":"10.1016\/j.neucom.2026.133466_bib1015","series-title":"ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","first-page":"1","article-title":"Gtn-Bailando: genre consistent long-term 3d dance generation based on pre-trained genre token network","author":"Zhuang","year":"2023"},{"key":"10.1016\/j.neucom.2026.133466_bib1020","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3485664","article-title":"Music2dance: Dancenet for music-driven dance generation","volume":"18","author":"Zhuang","year":"2022","journal-title":"ACM Trans. Multimed. Comput. Commun. Appl."},{"key":"10.1016\/j.neucom.2026.133466_bib1025","author":"Zhuang"},{"key":"10.1016\/j.neucom.2026.133466_bib1030","author":"Zhuang"}],"container-title":["Neurocomputing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0925231226008635?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0925231226008635?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T20:30:19Z","timestamp":1776889819000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0925231226008635"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,6]]},"references-count":206,"alternative-id":["S0925231226008635"],"URL":"https:\/\/doi.org\/10.1016\/j.neucom.2026.133466","relation":{},"ISSN":["0925-2312"],"issn-type":[{"value":"0925-2312","type":"print"}],"subject":[],"published":{"date-parts":[[2026,6]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Music-driven dance generation: A comprehensive review","name":"articletitle","label":"Article Title"},{"value":"Neurocomputing","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.neucom.2026.133466","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"133466"}}