{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,12]],"date-time":"2025-12-12T13:40:05Z","timestamp":1765546805980,"version":"3.41.0"},"reference-count":31,"publisher":"Association for Computing Machinery (ACM)","issue":"1","license":[{"start":{"date-parts":[[2021,10,18]],"date-time":"2021-10-18T00:00:00Z","timestamp":1634515200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100000266","name":"UK Engineering and Physical Sciences Research Council","doi-asserted-by":"crossref","award":["EP\/R02572X\/1"],"award-info":[{"award-number":["EP\/R02572X\/1"]}],"id":[{"id":"10.13039\/501100000266","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100018876","name":"National Centre for Nuclear Robotics","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100018876","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100011878","name":"Flemish Government","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100011878","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Centre for Artificial Intelligence, Robotics and Human-Machine Systems","award":["C82092"],"award-info":[{"award-number":["C82092"]}]},{"DOI":"10.13039\/501100008530","name":"European Regional Development Fund","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100008530","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["J. Hum.-Robot Interact."],"published-print":{"date-parts":[[2022,3,31]]},"abstract":"<jats:p>We explore the effectiveness of a dynamically processed incremental referring description system using under-specified ambiguous descriptions that are then built upon using linguistic repair statements, which we refer to as a dynamic system. We build a dynamically processed incremental referring description generation system that is able to provide contextual navigational statements to describe an object in a potential real-world situation of nuclear waste sorting and maintenance. In a study of 31 participants, we test the dynamic system in a case where a user is remote operating a robot to sort nuclear waste, with the robot assisting them in identifying the correct barrels to be removed. We compare these against a static non-ambiguous description given in the same scenario. As well as looking at efficiency with time and distance measurements, we also look at user preference. Results show that our dynamic system was a much more efficient method\u2014taking only 62% of the time on average\u2014for finding the correct barrel. Participants also favoured our dynamic system.<\/jats:p>","DOI":"10.1145\/3481628","type":"journal-article","created":{"date-parts":[[2021,10,19]],"date-time":"2021-10-19T00:28:21Z","timestamp":1634603301000},"page":"1-24","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":4,"title":["The Effectiveness of Dynamically Processed Incremental Descriptions in Human Robot Interaction"],"prefix":"10.1145","volume":"11","author":[{"given":"Christopher D.","family":"Wallbridge","sequence":"first","affiliation":[{"name":"IROHMS, University of Cardiff and University of Plymouth, Cardiff, Wales, UK"}]},{"given":"Alex","family":"Smith","sequence":"additional","affiliation":[{"name":"Bristol Robotics Laboratory, Coldharbour Ln, Bristol, UK"}]},{"given":"Manuel","family":"Giuliani","sequence":"additional","affiliation":[{"name":"Bristol Robotics Laboratory, Coldharbour Ln, Bristol, UK"}]},{"given":"Chris","family":"Melhuish","sequence":"additional","affiliation":[{"name":"Bristol Robotics Laboratory, Coldharbour Ln, Bristol, UK"}]},{"given":"Tony","family":"Belpaeme","sequence":"additional","affiliation":[{"name":"IDlab - imec, Ghent University and University of Plymouth, Ghent, Belgium"}]},{"given":"S\u00e9verin","family":"Lemaignan","sequence":"additional","affiliation":[{"name":"Bristol Robotics Laboratory, Coldharbour Ln, Bristol, UK"}]}],"member":"320","published-online":{"date-parts":[[2021,10,18]]},"reference":[{"key":"e_1_3_2_2_2","unstructured":"GOV.UK. 2020. Direct Research Portfolio Annual Report 2018 to 2019. Retrieved August 23 2021 from https:\/\/www.gov.uk\/government\/publications\/direct-research-portfolio-annual-report-2018-to-2019\/direct-research-annual-report-2018-to-2019."},{"key":"e_1_3_2_3_2","doi-asserted-by":"publisher","DOI":"10.1007\/s12369-008-0001-3"},{"key":"e_1_3_2_4_2","doi-asserted-by":"publisher","DOI":"10.1207\/s15516709cog1302_7"},{"key":"e_1_3_2_5_2","doi-asserted-by":"publisher","DOI":"10.1016\/0010-0277(86)90010-7"},{"key":"e_1_3_2_6_2","doi-asserted-by":"publisher","DOI":"10.1207\/s15516709cog1902_3"},{"key":"e_1_3_2_7_2","doi-asserted-by":"publisher","DOI":"10.3115\/1225753.1225754"},{"key":"e_1_3_2_8_2","doi-asserted-by":"publisher","DOI":"10.1007\/s11412-016-9235-5"},{"key":"e_1_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.1109\/IROS40897.2019.8968510"},{"key":"e_1_3_2_10_2","doi-asserted-by":"publisher","DOI":"10.1145\/2696454.2696467"},{"key":"e_1_3_2_11_2","doi-asserted-by":"publisher","DOI":"10.1163\/9789004368811_003"},{"key":"e_1_3_2_12_2","doi-asserted-by":"publisher","DOI":"10.5555\/216261.216267"},{"key":"e_1_3_2_13_2","doi-asserted-by":"publisher","DOI":"10.3115\/1220175.1220306"},{"key":"e_1_3_2_14_2","doi-asserted-by":"publisher","DOI":"10.1162\/COLI_a_00088"},{"key":"e_1_3_2_15_2","doi-asserted-by":"publisher","DOI":"10.1109\/ROMAN.2010.5598715"},{"key":"e_1_3_2_16_2","doi-asserted-by":"publisher","DOI":"10.1007\/s12369-011-0123-x"},{"key":"e_1_3_2_17_2","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2018.8594094"},{"key":"e_1_3_2_18_2","doi-asserted-by":"publisher","DOI":"10.5555\/2392800.2392827"},{"key":"e_1_3_2_19_2","first-page":"78","volume-title":"Proceedings of the SIGDIAL 2013 Conference","author":"Liu Changsong","year":"2013","unstructured":"Changsong Liu, Rui Fang, Lanbo She, and Joyce Chai. 2013. Modeling collaborative referring for situated referential grounding. In Proceedings of the SIGDIAL 2013 Conference. 78\u201386."},{"key":"e_1_3_2_20_2","doi-asserted-by":"crossref","unstructured":"Aly Magassouba Komei Sugiura and Hisashi Kawai. 2018. A multimodal classifier generative adversarial network for carry and place tasks from ambiguous language instructions. arXiv:1806.03847.","DOI":"10.1109\/LRA.2018.2849607"},{"key":"e_1_3_2_21_2","unstructured":"Aly Magassouba Komei Sugiura and Hisashi Kawai. 2019. Multimodal attention branch network for perspective-free sentence generation. arXiv:1909.05664."},{"key":"e_1_3_2_22_2","doi-asserted-by":"publisher","DOI":"10.5555\/1953048.2078195"},{"key":"e_1_3_2_23_2","doi-asserted-by":"publisher","DOI":"10.1007\/s11168-006-9004-0"},{"key":"e_1_3_2_24_2","doi-asserted-by":"publisher","DOI":"10.1109\/ROMAN.2010.5598719"},{"key":"e_1_3_2_25_2","doi-asserted-by":"crossref","unstructured":"Mohit Shridhar and David Hsu. 2018. Interactive visual grounding of referring expressions for human-robot interaction. arXiv:1806.03831.","DOI":"10.15607\/RSS.2018.XIV.028"},{"key":"e_1_3_2_26_2","volume-title":"Proceedings of the 25th AAAI Conference on Artificial Intelligence","author":"Tellex Stefanie","year":"2011","unstructured":"Stefanie Tellex, Thomas Kollar, Steven Dickerson, Matthew R. Walter, Ashis Gopal Banerjee, Seth Teller, and Nicholas Roy. 2011. Understanding natural language commands for robotic navigation and mobile manipulation. In Proceedings of the 25th AAAI Conference on Artificial Intelligence."},{"key":"e_1_3_2_27_2","unstructured":"James Temperton. 2017. Inside Sellafield: How the UK\u2019s most dangerous nuclear site is cleaning up its act. Wired . Retrieved August 23 2021 from https:\/\/www.wired.co.uk\/article\/inside-sellafield-nuclear-waste-decommissioning."},{"key":"e_1_3_2_28_2","doi-asserted-by":"publisher","DOI":"10.5555\/1708322.1708334"},{"key":"e_1_3_2_29_2","doi-asserted-by":"publisher","DOI":"10.3389\/frobt.2019.00067"},{"key":"e_1_3_2_30_2","volume-title":"Proceedings of the 4th Workshop on Robots for Learning (R4L): Inclusive Learning @HRI 2018","author":"Wallbridge Christopher David","year":"2018","unstructured":"Christopher David Wallbridge, S\u00e9verin Lemaignan, Emmanuel Senft, Charlotte Edmunds, and Tony Belpaeme. 2018. Spatial referring expressions in child-robot interaction: Let\u2019s be ambiguous! In Proceedings of the 4th Workshop on Robots for Learning (R4L): Inclusive Learning @HRI 2018."},{"key":"e_1_3_2_31_2","doi-asserted-by":"publisher","DOI":"10.1109\/HRI.2019.8673275"},{"key":"e_1_3_2_32_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W17-3511"}],"container-title":["ACM Transactions on Human-Robot Interaction"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3481628","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3481628","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:48:12Z","timestamp":1750193292000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3481628"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,10,18]]},"references-count":31,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2022,3,31]]}},"alternative-id":["10.1145\/3481628"],"URL":"https:\/\/doi.org\/10.1145\/3481628","relation":{},"ISSN":["2573-9522","2573-9522"],"issn-type":[{"type":"print","value":"2573-9522"},{"type":"electronic","value":"2573-9522"}],"subject":[],"published":{"date-parts":[[2021,10,18]]},"assertion":[{"value":"2020-01-01","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2021-07-01","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2021-10-18","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}