{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T20:15:48Z","timestamp":1776111348622,"version":"3.50.1"},"reference-count":40,"publisher":"Frontiers Media SA","license":[{"start":{"date-parts":[[2024,5,2]],"date-time":"2024-05-02T00:00:00Z","timestamp":1714608000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/100010244","name":"Colby College","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100010244","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["frontiersin.org"],"crossmark-restriction":true},"short-container-title":["Front. Robot. AI"],"abstract":"<jats:p>Dance plays a vital role in human societies across time and culture, with different communities having invented different systems for artistic expression through movement (genres). Differences between genres can be described by experts in words and movements, but these descriptions can only be appreciated by people with certain background abilities. Existing dance notation schemes could be applied to describe genre-differences, however they fall substantially short of being able to capture the important details of movement across a wide spectrum of genres. Our knowledge and practice around dance would benefit from a general, quantitative and human-understandable method of characterizing meaningful differences between aspects of any dance style; a computational kinematics of dance. Here we introduce and apply a novel system for encoding bodily movement as 17 macroscopic, interpretable features, such as expandedness of the body or the frequency of sharp movements. We use this encoding to analyze Hip Hop Dance genres, in part by building a low-cost machine-learning classifier that distinguishes genre with high accuracy. Our study relies on an open dataset (AIST++) of pose-sequences from dancers instructed to perform one of ten Hip Hop genres, such as Breakdance, Popping, or Krump. For comparison we evaluate moderately experienced human observers at discerning these sequence\u2019s genres from movements alone (38% where chance = 10%). The performance of a baseline, Ridge classifier model was fair (48%) and that of the model resulting from our automated machine learning pipeline was strong (76%). This indicates that the selected features represent important dimensions of movement for the expression of the attitudes, stories, and aesthetic values manifested in these dance forms. Our study offers a new window into significant relations of similarity and difference between the genres studied. Given the rich, complex, and culturally shaped nature of these genres, the interpretability of our features, and the lightweight techniques used, our approach has significant potential for generalization to other movement domains and movement-related applications.<\/jats:p>","DOI":"10.3389\/frobt.2024.1295308","type":"journal-article","created":{"date-parts":[[2024,5,2]],"date-time":"2024-05-02T15:29:15Z","timestamp":1714663755000},"update-policy":"https:\/\/doi.org\/10.3389\/crossmark-policy","source":"Crossref","is-referenced-by-count":12,"title":["Computational kinematics of dance: distinguishing hip hop genres"],"prefix":"10.3389","volume":"11","author":[{"given":"Ben","family":"Baker","sequence":"first","affiliation":[]},{"given":"Tony","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Jordan","family":"Matelsky","sequence":"additional","affiliation":[]},{"given":"Felipe","family":"Parodi","sequence":"additional","affiliation":[]},{"given":"Brett","family":"Mensh","sequence":"additional","affiliation":[]},{"given":"John W.","family":"Krakauer","sequence":"additional","affiliation":[]},{"given":"Konrad","family":"Kording","sequence":"additional","affiliation":[]}],"member":"1965","published-online":{"date-parts":[[2024,5,2]]},"reference":[{"key":"B1","doi-asserted-by":"crossref","first-page":"171","DOI":"10.1093\/oxfordhb\/9780190247867.013.29","article-title":"Framing hip hop dance as an object of sociological and cultural research","volume-title":"The oxford handbook of hip hop dance studies","author":"Bennett","year":"2022"},{"key":"B2","doi-asserted-by":"publisher","first-page":"411","DOI":"10.1007\/bf03395235","article-title":"Perception of movement and dancer characteristics from point-light displays of dance","volume":"47","author":"Brownlow","year":"1997","journal-title":"Psychol. Rec."},{"key":"B3","doi-asserted-by":"publisher","first-page":"1243","DOI":"10.1093\/cercor\/bhi007","article-title":"Action observation and acquired motor skills: an FMRI study with expert dancers","volume":"15","author":"Calvo-Merino","year":"2005","journal-title":"Cereb. Cortex"},{"key":"B4","doi-asserted-by":"publisher","first-page":"162","DOI":"10.1080\/09298215.2020.1711778","article-title":"Dance to your own drum: identification of musical genre and individual dancer from motion capture using machine learning","volume":"49","author":"Carlson","year":"2020","journal-title":"J. New Music Res."},{"key":"B5","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1801.07388","article-title":"Let\u2019s dance: learning from online dance videos","author":"Castro","year":"2018","journal-title":"arXiv [cs.CV]"},{"key":"B6","volume-title":"Can\u2019t stop, won't stop: a history of the hip-hop generation","author":"Chang","year":"2005"},{"key":"B7","volume-title":"The big payback the history of the business of hip-hop","author":"Charnas","year":"2010"},{"key":"B8","doi-asserted-by":"crossref","first-page":"243","DOI":"10.1093\/oxfordhb\/9780190247867.013.3","article-title":"Breaking in my House: popular dance, gender identities, and postracial empathies","volume-title":"The oxford handbook of hip hop dance studies","author":"DeFrantz","year":"2022"},{"key":"B9","doi-asserted-by":"publisher","first-page":"727","DOI":"10.1068\/p250727","article-title":"Perception of emotion from dynamic point-light displays represented in dance","volume":"25","author":"Dittrich","year":"1996","journal-title":"Perception"},{"key":"B10","doi-asserted-by":"crossref","DOI":"10.1109\/CVPR52688.2022.00298","article-title":"Revisiting skeleton-based action recognition","volume-title":"2022 IEEE\/CVF conference on computer vision and pattern recognition (CVPR)","author":"Duan","year":"2022"},{"key":"B11","doi-asserted-by":"crossref","first-page":"80","DOI":"10.1093\/oxfordhb\/9780190247867.013.4","article-title":"Connecting hip hop history and heritage","volume-title":"The oxford handbook of hip hop dance studies","author":"Durden","year":"2022"},{"key":"B12","first-page":"192","article-title":"Dance in the world of data and objects","volume-title":"Information technologies for performing arts, media access, and entertainment","author":"El","year":"2013"},{"key":"B13","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2007.04074","article-title":"Auto-sklearn 2.0: hands-free AutoML via meta-learning","author":"Feurer","year":"2020","journal-title":"arXiv [cs.LG]"},{"key":"B14","article-title":"Efficient and robust automated machine learning","volume-title":"Advances in neural information processing systems","author":"Feurer","year":"2015"},{"key":"B15","doi-asserted-by":"crossref","DOI":"10.5744\/florida\/9780813049298.001.0001","volume-title":"Jazz dance: a history of the roots and branches","author":"Guarino","year":"2014"},{"key":"B16","volume-title":"Labanotation: or, kinetography laban: the system of analyzing and recording movement","author":"Guest","year":"1970"},{"key":"B17","doi-asserted-by":"publisher","first-page":"456","DOI":"10.1016\/j.neubiorev.2020.09.036","article-title":"I tried a bunch of things: the dangers of unexpected overfitting in classification of brain data","volume":"119","author":"Hosseini","year":"2020","journal-title":"Neurosci. Biobehav. Rev."},{"key":"B18","doi-asserted-by":"publisher","first-page":"201","DOI":"10.3758\/bf03212378","article-title":"Visual perception of biological motion and a model for its analysis","volume":"14","author":"Johansson","year":"1973","journal-title":"Percept. Psychophys."},{"key":"B19","doi-asserted-by":"crossref","DOI":"10.1109\/MRA.2011.942118","article-title":"Automatic sequencing of ballet poses","volume-title":"IEEE robotics and automation magazine\/IEEE robotics and automation society","author":"LaViers","year":"2011"},{"key":"B20","volume-title":"AI choreographer: music conditioned 3D dance generation with AIST++","author":"Li","year":"2021"},{"key":"B21","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2816795.2818013","article-title":"SMPL: a skinned multi-person linear model","volume":"248","author":"Loper","year":"2015","journal-title":"ACM Trans. Graph."},{"key":"B22","volume-title":"A unified approach to interpreting model predictions","author":"Lundberg","year":"2017"},{"key":"B23","doi-asserted-by":"publisher","first-page":"176","DOI":"10.1162\/daed_a_00086","article-title":"Hip-hop and the global imprint of a Black cultural form","volume":"140","author":"Morgan","year":"2011","journal-title":"Daedalus"},{"key":"B24","volume-title":"New cultural studies of dance","year":"1997"},{"key":"B25","doi-asserted-by":"crossref","DOI":"10.1007\/978-1-137-05964-2","volume-title":"The africanist aesthetic in global hip-hop","author":"Osumare","year":"2007"},{"key":"B26","doi-asserted-by":"crossref","DOI":"10.1145\/3212721.3212809","article-title":"A design methodology for abstracting character archetypes onto robotic systems","volume-title":"Proceedings of the 5th international","author":"Pakrasi","year":"2018"},{"key":"B27","doi-asserted-by":"publisher","first-page":"2259","DOI":"10.1007\/s10462-020-09904-8","article-title":"A survey on video-based human action recognition: recent updates, datasets, challenges, and applications","volume":"54","author":"Pareek","year":"2021","journal-title":"Artif. Intell. Rev."},{"key":"B28","doi-asserted-by":"crossref","DOI":"10.1057\/9781403981677","volume-title":"New York ricans from the hip hop zone","author":"Rivera","year":"2003"},{"key":"B29","doi-asserted-by":"crossref","DOI":"10.1101\/059774","volume-title":"Voodoo machine learning for clinical predictions","author":"Saeb","year":"2016"},{"key":"B30","doi-asserted-by":"publisher","first-page":"1627","DOI":"10.1021\/ac60214a047","article-title":"Smoothing and differentiation of data by simplified least squares procedures","volume":"36","author":"Savitzky","year":"1964","journal-title":"Anal. Chem."},{"key":"B31","volume-title":"Foundation: B-boys, B-girls and hip-hop culture in New York","author":"Schloss","year":"2009"},{"key":"B32","doi-asserted-by":"publisher","first-page":"447","DOI":"10.1007\/s00221-012-3229-y","article-title":"Perceiving bodies in motion: expression intensity, empathy, and experience","volume":"222","author":"Sevdalis","year":"2012","journal-title":"Exp. Brain Res. Exp. Hirnforschung. Exp. Cerebrale"},{"key":"B33","doi-asserted-by":"publisher","first-page":"2166","DOI":"10.1039\/d0lc00096e","article-title":"A web-based automated machine learning platform to analyze liquid biopsy data","volume":"20","author":"Shen","year":"2020","journal-title":"Lab a Chip"},{"key":"B34","volume-title":"Bailando: 3D dance generation by actor-critic GPT with choreographic memory","author":"Siyao","year":"2022"},{"key":"B35","doi-asserted-by":"crossref","DOI":"10.1007\/978-1-137-48777-3","volume-title":"The body, dance and cultural theory","author":"Thomas","year":"2003"},{"key":"B36","volume-title":"EDGE: editable dance generation from music","author":"Tseng","year":"2022"},{"key":"B37","unstructured":"TsuchidaS.\n          2023"},{"key":"B38","doi-asserted-by":"publisher","first-page":"53880","DOI":"10.1109\/access.2023.3282311","article-title":"A comprehensive survey of RGB-based and skeleton-based human action recognition","volume":"11","author":"Wang","year":"2023","journal-title":"IEEE Access"},{"key":"B39","first-page":"103225","volume-title":"Deep 3D human pose estimation: a review.\u201d computer vision and image understanding: CVIU 210 (september)","author":"Wang","year":"2021"},{"key":"B40","first-page":"7444","article-title":"Spatial temporal graph convolutional networks for skeleton-based action recognition","volume-title":"Proceedings of the thirty-second AAAI conference on artificial intelligence and thirtieth innovative applications of artificial intelligence conference and eighth AAAI symposium on educational advances in artificial intelligence","author":"Yan","year":"2018"}],"container-title":["Frontiers in Robotics and AI"],"original-title":[],"link":[{"URL":"https:\/\/www.frontiersin.org\/articles\/10.3389\/frobt.2024.1295308\/full","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,5,2]],"date-time":"2024-05-02T15:29:34Z","timestamp":1714663774000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.frontiersin.org\/articles\/10.3389\/frobt.2024.1295308\/full"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,2]]},"references-count":40,"alternative-id":["10.3389\/frobt.2024.1295308"],"URL":"https:\/\/doi.org\/10.3389\/frobt.2024.1295308","relation":{},"ISSN":["2296-9144"],"issn-type":[{"value":"2296-9144","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,5,2]]},"article-number":"1295308"}}