{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,23]],"date-time":"2025-06-23T16:10:10Z","timestamp":1750695010312,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":20,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,6,23]]},"DOI":"10.1145\/3698061.3734415","type":"proceedings-article","created":{"date-parts":[[2025,6,23]],"date-time":"2025-06-23T06:55:33Z","timestamp":1750661733000},"page":"491-493","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Interactive Movement-to-Audio with Pre-Trained Neural Networks"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-6314-1635","authenticated-orcid":false,"given":"Joseph","family":"Meyer","sequence":"first","affiliation":[{"name":"Creative Computing Institute, University of the Arts London, London, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1382-2914","authenticated-orcid":false,"given":"Nick","family":"Bryan-Kinns","sequence":"additional","affiliation":[{"name":"Creative Computing Institute, University of the Arts London, London, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1775-213X","authenticated-orcid":false,"given":"Sarah","family":"Fdili Alaoui","sequence":"additional","affiliation":[{"name":"Creative Computing Institute, University of the Arts London, London, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6981-5414","authenticated-orcid":false,"given":"Mick","family":"Grierson","sequence":"additional","affiliation":[{"name":"Creative Computing Institute, University of the Arts London, London, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7609-2234","authenticated-orcid":false,"given":"Rebecca","family":"Fiebrink","sequence":"additional","affiliation":[{"name":"Creative Computing Institute, University of the Arts London, London, United Kingdom"}]}],"member":"320","published-online":{"date-parts":[[2025,6,22]]},"reference":[{"key":"e_1_3_3_2_1_2","unstructured":"Rishabh Bajpai and Deepak Joshi. 2021. MoveNet: A deep neural network for joint profile prediction across variable walking speeds and slopes. https:\/\/ieeexplore.ieee.org\/document\/9406043"},{"key":"e_1_3_3_2_2_2","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3581488"},{"key":"e_1_3_3_2_3_2","volume-title":"Gesture Capture: Paradigms in Interactive Music\/Dance System. https:\/\/www.researchgate.net\/publication\/267946741_Gesture_Capture_Paradigms_in_Interactive_Music_Dance_Systems","author":"Bevilacqua Fr\u00e9d\u00e9ric","year":"2011","unstructured":"Fr\u00e9d\u00e9ric Bevilacqua, Norbert Schnell, and Sarah Fdili Alaoui. 2011. Gesture Capture: Paradigms in Interactive Music\/Dance System. https:\/\/www.researchgate.net\/publication\/267946741_Gesture_Capture_Paradigms_in_Interactive_Music_Dance_Systems"},{"key":"e_1_3_3_2_4_2","volume-title":"Proceedings of Explainable AI for the Arts Workshop 2024 (XAIxArts 2024)","author":"Bryan-Kinns Nick","year":"2024","unstructured":"Nick Bryan-Kinns and Zijin Li. 2024. Reducing Barriers to the Use of Marginalised Music Genres in AI. In Proceedings of Explainable AI for the Arts Workshop 2024 (XAIxArts 2024) at ACM Creativity and Cognition 2023. https:\/\/arxiv.org\/pdf\/2407.13439"},{"key":"e_1_3_3_2_5_2","volume-title":"RAVE: A variational autoencoder for fast and high-quality neural audio synthesis. https:\/\/arxiv.org\/pdf\/2111.05011.pdf","author":"Caillon Antoine","year":"2021","unstructured":"Antoine Caillon and Philippe Esling. 2021. RAVE: A variational autoencoder for fast and high-quality neural audio synthesis. https:\/\/arxiv.org\/pdf\/2111.05011.pdf"},{"key":"e_1_3_3_2_6_2","unstructured":"Zhe Cao et al. 2019. OpenPose: Realtime multi-person 2d pose estimation using part affinity fields. https:\/\/arxiv.org\/pdf\/1812.08008.pdf"},{"volume-title":"International Conference on Machine Learning (pp. 1068-1077)","author":"Jesse","key":"e_1_3_3_2_7_2","unstructured":"Jesse Engel et al. 2017. Neural audio synthesis of musical notes with wavenet autoencoders. In International Conference on Machine Learning (pp. 1068-1077). https:\/\/arxiv.org\/pdf\/1704.01279"},{"key":"e_1_3_3_2_8_2","unstructured":"Rebecca Fiebrink et al. 2009. A Meta-Instrument for Interactive On-the-fly Machine Learning. In New Interfaces for Musical Expression 2009. https:\/\/www.cs.princeton.edu\/sound\/publications\/FiebrinkTruemanCook_NIME2009.pdf"},{"key":"e_1_3_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.1145\/3491102.3501916"},{"key":"e_1_3_3_2_10_2","doi-asserted-by":"publisher","unstructured":"Andy Hunt and Marcelo Wanderley. 2002. Mapping performer parameters to synthesis engines. 10.1017\/S1355771802002030","DOI":"10.1017\/S1355771802002030"},{"key":"e_1_3_3_2_11_2","doi-asserted-by":"publisher","unstructured":"BeomJun Jo and SeongKi Kim. 2022. Comparative Analysis of OpenPose PoseNet and MoveNet Models for Pose Estimation in Mobile Devices. 10.18280\/ts.390111","DOI":"10.18280\/ts.390111"},{"key":"e_1_3_3_2_12_2","unstructured":"Thomas Mitchell et al. 2012. Musical Interaction with Hand Posture and Orientation: A Toolbox of Gestural Control Mechanisms. https:\/\/www.nime.org\/proceedings\/2012\/nime2012_272.pdf"},{"key":"e_1_3_3_2_13_2","doi-asserted-by":"publisher","unstructured":"Sarah Nabi et al. 2024. Embodied exploration of deep latent spaces in interactive dance-music performance. 10.1145\/3658852.3659072","DOI":"10.1145\/3658852.3659072"},{"key":"e_1_3_3_2_14_2","volume-title":"Wavenet: A generative model for raw audio. https:\/\/arxiv.org\/pdf\/1609.03499.pdf","author":"van den Oord A\u00e4ron","year":"2016","unstructured":"A\u00e4ron van den Oord et al. 2016. Wavenet: A generative model for raw audio. https:\/\/arxiv.org\/pdf\/1609.03499.pdf"},{"key":"e_1_3_3_2_15_2","doi-asserted-by":"crossref","unstructured":"George Papandreou et al. 2018. PersonLab: Person pose estimation and instance segmentation with a bottom-up part-based geometric embedding model. https:\/\/arxiv.org\/pdf\/1803.08225","DOI":"10.1007\/978-3-030-01264-9_17"},{"key":"e_1_3_3_2_16_2","doi-asserted-by":"publisher","unstructured":"Mitchel Resnick et al. 2005. Design Principles for Tools to Support Creative Thinking. 10.1184\/R1\/6621917.v1","DOI":"10.1184\/R1\/6621917.v1"},{"volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (pp. 10219-10228)","author":"Ludan","key":"e_1_3_3_2_17_2","unstructured":"Ludan Ruan et al. 2023. Mm-diffusion: Learning multi-modal diffusion models for joint audio and video generation. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (pp. 10219-10228)."},{"key":"e_1_3_3_2_18_2","unstructured":"Stefania Serafin et al. 2014. Controlling Physically Based Virtual Musical Instruments Using The Gloves. https:\/\/www.nime.org\/proceedings\/2014\/nime2014_307.pdf"},{"key":"e_1_3_3_2_19_2","doi-asserted-by":"crossref","unstructured":"Gabriel Vigliensoni and Rebecca Fiebrink. 2023. Steering latent audio models through interactive machine learning. https:\/\/ualresearchonline.arts.ac.uk\/id\/eprint\/20199\/1\/VigliensoniFiebrink_ICCC2023.pdf","DOI":"10.5920\/jcms.902"},{"key":"e_1_3_3_2_20_2","unstructured":"Shuoyang Zheng et al. 2024. A Mapping Strategy for Interacting with Latent Audio Synthesis Using Artistic Materials. https:\/\/arxiv.org\/pdf\/2407.04379"}],"event":{"name":"C&C '25: Creativity and Cognition","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction"],"location":"Virtual United Kingdom","acronym":"C&C '25"},"container-title":["Proceedings of the 2025 Conference on Creativity and Cognition"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3698061.3734415","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,23]],"date-time":"2025-06-23T15:52:53Z","timestamp":1750693973000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3698061.3734415"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,22]]},"references-count":20,"alternative-id":["10.1145\/3698061.3734415","10.1145\/3698061"],"URL":"https:\/\/doi.org\/10.1145\/3698061.3734415","relation":{},"subject":[],"published":{"date-parts":[[2025,6,22]]},"assertion":[{"value":"2025-06-22","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}