{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,4]],"date-time":"2025-11-04T06:04:25Z","timestamp":1762236265251,"version":"build-2065373602"},"reference-count":57,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,8,25]],"date-time":"2025-08-25T00:00:00Z","timestamp":1756080000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,8,25]],"date-time":"2025-08-25T00:00:00Z","timestamp":1756080000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,8,25]]},"DOI":"10.1109\/ro-man63969.2025.11217907","type":"proceedings-article","created":{"date-parts":[[2025,11,3]],"date-time":"2025-11-03T18:42:29Z","timestamp":1762195349000},"page":"199-206","source":"Crossref","is-referenced-by-count":0,"title":["Do Re Mi Fa So Pass the Tool: Using Melodic Prediction to Improve Human-Robot Fluency"],"prefix":"10.1109","author":[{"given":"Amit","family":"Rogel","sequence":"first","affiliation":[{"name":"Robotic Musicianship Lab Georgia Tech,Atlanta,GA,USA"}]},{"given":"Qiaoyu","family":"Yang","sequence":"additional","affiliation":[{"name":"Robotic Musicianship Lab Georgia Tech,Atlanta,GA,USA"}]},{"given":"Jack","family":"Hayley","sequence":"additional","affiliation":[{"name":"Robotic Musicianship Lab Georgia Tech,Atlanta,GA,USA"}]},{"given":"Gil","family":"Weinberg","sequence":"additional","affiliation":[{"name":"Robotic Musicianship Lab Georgia Tech,Atlanta,GA,USA"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/thms.2019.2904558"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1561\/2300000052"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1016\/S0921-8890(02)00373-1"},{"key":"ref4","first-page":"1259","article-title":"Impact of alarm fatigue on the work of nurses in an intensive care environment\u2014a systematic review","volume":"35","author":"Konkani","year":"2021","journal-title":"Journal of Clinical Monitoring and Computing"},{"key":"ref5","article-title":"To reduce hospital noise, researchers create alarms that whistle and sing","author":"Rueb","year":"2019","journal-title":"The New York Times"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.2345\/0899-8205-46.4.268"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1177\/0018720817712004"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1093\/acprof:oso\/9780198529361.003.0008"},{"key":"ref9","first-page":"189","article-title":"Resonating to musical rhythm: Theory and experiment","volume-title":"Psychology of Time","author":"Large","year":"2008"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1525\/mp.2010.27.4.263"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.7551\/mitpress\/6575.001.0001"},{"issue":"2","key":"ref12","first-page":"315","article-title":"Melodic expectation: A review and analysis","volume":"51","author":"Margulis","year":"2007","journal-title":"Journal of Music Theory"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.7208\/chicago\/9780226521374.001.0001"},{"article-title":"What is a leading tone?","year":"2023","author":"Hein","key":"ref14"},{"volume-title":"The analysis and cognition of basic melodic structures: The implication-realization model","year":"1990","author":"Narmour","key":"ref15"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.2979\/6363.0"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.2307\/40285249"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.2307\/40285384"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1525\/mp.2005.22.4.663"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1093\/acprof:oso\/9780199230143.001.0001"},{"issue":"1","key":"ref21","first-page":"65","article-title":"Effects of anticipatory action on human-robot teamwork efficiency, fluency, and perception of team","volume":"30","author":"Hoffman","year":"2011","journal-title":"Autonomous Robots"},{"key":"ref22","first-page":"137 098","article-title":"Deep motion prediction for human-robot interaction using bi-directional lstm","volume":"8","author":"Zhao","year":"2020","journal-title":"IEEE Access"},{"key":"ref23","first-page":"26","article-title":"Vision-based grasping for object manipulation in human-robot interaction","volume":"5","author":"Ciliberto","year":"2018","journal-title":"Frontiers in Robotics and AI"},{"key":"ref24","first-page":"51","article-title":"Effects of robot motion on human-robot collaboration","volume-title":"Proceedings of the 10th Annual ACM\/IEEE International Conference on Human-Robot Interaction","author":"Dragan"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2001.976403"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-50115-4_30"},{"key":"ref27","first-page":"464","article-title":"Jam\u2019aa-a middle eastern percussion ensemble for human and robotic players","author":"Weinberg","year":"2006","journal-title":"Icmc"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1145\/3537972.3537985"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/HRI53351.2022.9889384"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1201\/9781003320470-13"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/MRA.2022.3210565"},{"key":"ref32","first-page":"3685","article-title":"Augmented reality visual indicators to improve shared experiences in human-robot collaboration","volume-title":"IEEE\/RSJ International Conference on Intelligent Robots and Systems","author":"Rosen"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1016\/j.jmsy.2024.05.001"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.3389\/fphys.2017.00785"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0070758"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/ROMAN.2014.6926327"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1097\/aco.0000000000000260"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1016\/j.pcorm.2023.100332"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1016\/j.bja.2022.10.045"},{"key":"ref40","first-page":"205","article-title":"Joint action in music performance","volume-title":"Emerging Communication","volume":"10","author":"Keller","year":"2008"},{"article-title":"Sonification in industrial robotics: enhancing worker safety through auditory cues","volume-title":"International Conference on Industry 4.0 and Smart Manufacturing","author":"Haynes","key":"ref41"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1155\/2013\/586138"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2022.3193228"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1016\/j.apergo.2020.103068"},{"volume-title":"Musical robots and interactive multimodal systems","year":"2010","author":"Hoffman","key":"ref45"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/RO-MAN47096.2020.9223452"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1038\/srep42005"},{"article-title":"Sonification from a design perspective","volume-title":"Proceedings of the 9th International Conference on Auditory Display","author":"Barrass","key":"ref48"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1007\/s10804-010-9117-4"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1525\/mp.2020.37.4.298"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.3389\/fpsyg.2015.01140"},{"issue":"11","key":"ref52","first-page":"905","article-title":"Complex auditory displays in healthcare environments","volume":"67","author":"Sanderson","year":"2009","journal-title":"International Journal of Human-Computer Studies"},{"issue":"31","key":"ref53","first-page":"210","article-title":"Auditory warnings in noisy environments","volume":"8","author":"Edworthy","year":"2006","journal-title":"Noise & Health"},{"key":"ref54","first-page":"1572","article-title":"Effects of Auditory Cues on Human-Robot Collaboration","volume-title":"2020 IEEE 29th International Symposium on Industrial Electronics (ISIE)","author":"Okimoto"},{"key":"ref55","first-page":"218","article-title":"Spatial auditory cues improve human-robot interaction: evidence from a dual-task paradigm","volume-title":"Proceedings of the ACM\/IEEE International Conference on Human-Robot Interaction","author":"Huang"},{"issue":"4","key":"ref56","first-page":"573","article-title":"Auditory-motor entrainment and spatiotemporal coordination","volume":"80","author":"Van der Wel","year":"2016","journal-title":"Psychological Research"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0110274"}],"event":{"name":"2025 34th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN)","start":{"date-parts":[[2025,8,25]]},"location":"Eindhoven, Netherlands","end":{"date-parts":[[2025,8,29]]}},"container-title":["2025 34th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11217544\/11217526\/11217907.pdf?arnumber=11217907","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,4]],"date-time":"2025-11-04T06:01:55Z","timestamp":1762236115000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11217907\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,25]]},"references-count":57,"URL":"https:\/\/doi.org\/10.1109\/ro-man63969.2025.11217907","relation":{},"subject":[],"published":{"date-parts":[[2025,8,25]]}}}