{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,5]],"date-time":"2026-05-05T03:47:07Z","timestamp":1777952827153,"version":"3.51.4"},"reference-count":34,"publisher":"Frontiers Media SA","license":[{"start":{"date-parts":[[2023,7,28]],"date-time":"2023-07-28T00:00:00Z","timestamp":1690502400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["frontiersin.org"],"crossmark-restriction":true},"short-container-title":["Front. Robot. AI"],"abstract":"<jats:p>Robots currently provide only a limited amount of information about their future movements to human collaborators. In human interaction, communication through gaze can be helpful by intuitively directing attention to specific targets. Whether and how this mechanism could benefit the interaction with robots and how a design of predictive robot eyes in general should look like is not well understood. In a between-subjects design, four different types of eyes were therefore compared with regard to their attention directing potential: a pair of arrows, human eyes, and two anthropomorphic robot eye designs. For this purpose, 39 subjects performed a novel, screen-based gaze cueing task in the laboratory. Participants\u2019 attention was measured using manual responses and eye-tracking. Information on the perception of the tested cues was provided through additional subjective measures. All eye models were overall easy to read and were able to direct participants\u2019 attention. The anthropomorphic robot eyes were most efficient at shifting participants\u2019 attention which was revealed by faster manual and saccadic reaction times. In addition, a robot equipped with anthropomorphic eyes was perceived as being more competent. Abstract anthropomorphic robot eyes therefore seem to trigger a reflexive reallocation of attention. This points to a social and automatic processing of such artificial stimuli.<\/jats:p>","DOI":"10.3389\/frobt.2023.1178433","type":"journal-article","created":{"date-parts":[[2023,7,28]],"date-time":"2023-07-28T11:20:30Z","timestamp":1690543230000},"update-policy":"https:\/\/doi.org\/10.3389\/crossmark-policy","source":"Crossref","is-referenced-by-count":10,"title":["The potential of robot eyes as predictive cues in HRI\u2014an eye-tracking study"],"prefix":"10.3389","volume":"10","author":[{"given":"Linda","family":"Onnasch","sequence":"first","affiliation":[]},{"given":"Paul","family":"Schweidler","sequence":"additional","affiliation":[]},{"given":"Helena","family":"Schmidt","sequence":"additional","affiliation":[]}],"member":"1965","published-online":{"date-parts":[[2023,7,28]]},"reference":[{"key":"B1","first-page":"1983","article-title":"Robot gaze does not reflexively cue human attention","volume":"33","author":"Admoni","year":"2011","journal-title":"Proc. Annu. Meet. Cognitive Sci. Soc."},{"key":"B2","doi-asserted-by":"publisher","first-page":"25","DOI":"10.5898\/jhri.6.1.admoni","article-title":"Social eye gaze in human-robot interaction: A review","volume":"6","author":"Admoni","year":"2017","journal-title":"J. Human-Robot Interact."},{"key":"B3","first-page":"119","article-title":"Legible light communications for factory robots","author":"Bacula","year":"2020"},{"key":"B4","doi-asserted-by":"publisher","first-page":"1236","DOI":"10.3758\/s13414-019-01669-9","article-title":"The role of eye movements in manual responses to social and nonsocial cues","volume":"81","author":"Bonmassar","year":"2019","journal-title":"Atten. Percept. Psychophys."},{"key":"B5","doi-asserted-by":"publisher","first-page":"3","DOI":"10.3389\/fnbot.2012.00003","article-title":"I reach faster when i see you look: Gaze effects in human-human and human-robot face-to-face cooperation","volume":"6","author":"Boucher","year":"2012","journal-title":"Front. Neurorobotics"},{"key":"B6","first-page":"254","article-title":"The robotic social attributes scale (RoSAS) development and validation","author":"Carpinella","year":"2017"},{"key":"B7","doi-asserted-by":"publisher","first-page":"12","DOI":"10.3389\/fnbot.2013.00012","article-title":"Comparing the effect of humanoid and human face for the spatial orientation of attention","volume":"7","author":"Chaminade","year":"2013","journal-title":"Front. Neurorobotics"},{"key":"B36","first-page":"1031","article-title":"Understanding robots: Making robots more legible in multi-party interactions","author":"Faria","year":"2021"},{"key":"B9","doi-asserted-by":"publisher","first-page":"1149","DOI":"10.3758\/brm.41.4.1149","article-title":"Statistical power analyses using G*Power 3.1: Tests for correlation and regression analyses","volume":"41","author":"Faul","year":"2009","journal-title":"Behav. Res. Methods"},{"key":"B10","doi-asserted-by":"publisher","first-page":"175","DOI":"10.3758\/bf03193146","article-title":"G*Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences","volume":"39","author":"Faul","year":"2007","journal-title":"Behav. Res. Methods"},{"key":"B11","doi-asserted-by":"publisher","first-page":"490","DOI":"10.3758\/bf03208827","article-title":"The eyes have it! Reflexive orienting is triggered by nonpredictive gaze","volume":"5","author":"Friesen","year":"1998","journal-title":"Psychonomic Bull. Rev."},{"key":"B12","doi-asserted-by":"publisher","first-page":"319","DOI":"10.1037\/0096-1523.30.2.319","article-title":"Attentional effects of counterpredictive gaze and arrow cues","volume":"30","author":"Friesen","year":"2004","journal-title":"J. Exp. Psychol. Hum. Percept. Perform."},{"key":"B13","doi-asserted-by":"publisher","first-page":"5010","DOI":"10.1109\/lra.2021.3068708","article-title":"Hey robot, which way are you going? Nonverbal motion legibility cues for human-robot spatial interaction","volume":"6","author":"Hetherington","year":"2021","journal-title":"IEEE Robotics Automation Lett."},{"key":"B15","doi-asserted-by":"publisher","first-page":"01568744","DOI":"10.1371\/journal.pone.0156874","article-title":"Role of gaze cues in interpersonal motor coordination: Towards higher affiliation in human-robot interaction","volume":"11","author":"Khoramshahi","year":"2016","journal-title":"PLOS ONE"},{"key":"B16","first-page":"334","article-title":"Meet me where I\u2019m gazing: How shared attention gaze affects human-robot handover timing","author":"Moon","year":"2014"},{"key":"B17","first-page":"69","article-title":"Nonverbal leakage in robots: Communication of intentions through seemingly unintentional behavior","author":"Mutlu","year":"2009"},{"key":"B18","doi-asserted-by":"publisher","first-page":"718","DOI":"10.3758\/pbr.17.5.718","article-title":"Coordinating spatial referencing using shared gaze","volume":"17","author":"Neider","year":"2010","journal-title":"Psychonomic Bull. Rev."},{"key":"B19","doi-asserted-by":"publisher","first-page":"188","DOI":"10.3758\/brm.42.1.188","article-title":"An adaptive algorithm for fixation, saccade, and glissade detection in eyetracking data","volume":"42","author":"Nystr\u00f6m","year":"2010","journal-title":"Behav. Res. methods"},{"key":"B20","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3472224","article-title":"Impact of anthropomorphic robot design on trust and attention in industrial human-robot interaction","volume":"11","author":"Onnasch","year":"2021","journal-title":"ACM Trans. Human-Robot Interact. (THRI)"},{"key":"B21","doi-asserted-by":"publisher","first-page":"848295","DOI":"10.3389\/frobt.2022.848295","article-title":"Humans can\u2019t resist robot eyes \u2013 reflexive cueing with pseudo-social stimuli","volume":"9","author":"Onnasch","year":"2022","journal-title":"Front. Robotics AI"},{"key":"B22","doi-asserted-by":"publisher","first-page":"833","DOI":"10.1007\/s12369-020-00666-5","article-title":"A taxonomy to structure and analyze human\u2013robot interaction","volume":"13","author":"Onnasch","year":"2021","journal-title":"Int. J. Soc. Robotics"},{"key":"B23","first-page":"203","article-title":"Design of robot eyes suitable for gaze communication","author":"Onuki","year":"2013"},{"key":"B24","first-page":"152","article-title":"Joint action with icub: A successful adaptation of a paradigm of cognitive neuroscience in hri","author":"P\u00e9rez-Osorio","year":"2018"},{"key":"B25","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1080\/00335558008248231","article-title":"Orienting of attention","volume":"32","author":"Posner","year":"1980","journal-title":"Q. J. Exp. Psychol."},{"key":"B26","first-page":"219","article-title":"Iterative virtual guides programming for human-robot comanipulation","author":"Restrepo","year":"2017"},{"key":"B27","doi-asserted-by":"publisher","first-page":"2259","DOI":"10.1097\/00001756-200212030-00018","article-title":"My eyes want to look where your eyes are looking: Exploring the tendency to imitate another individual\u2019s gaze","volume":"13","author":"Ricciardelli","year":"2002","journal-title":"NeuroReport"},{"key":"B28","doi-asserted-by":"publisher","first-page":"B55","DOI":"10.1016\/j.cognition.2004.04.005","article-title":"Taking control of reflexive social attention","volume":"94","author":"Ristic","year":"2005","journal-title":"Cognition"},{"key":"B29","doi-asserted-by":"publisher","first-page":"5","DOI":"10.3389\/fnint.2010.00005","article-title":"Following gaze: Gaze-following behavior as a window into social cognition","volume":"4","author":"Shepherd","year":"2010","journal-title":"Front. Integr. Neurosci."},{"key":"B30","first-page":"1651","article-title":"Exploring the use of light and display indicators for communicating directional intent","author":"Shrestha","year":"2016"},{"key":"B31","article-title":"Exploring the effects of gaze and pauses in situated human-robot interaction","author":"Skantze","year":"2013"},{"key":"B32","first-page":"77","article-title":"Visual attention in spoken human-robot interaction","author":"Staudte","year":"2008"},{"key":"B33","doi-asserted-by":"publisher","first-page":"879","DOI":"10.1162\/jocn.2008.20510","article-title":"Brain responses to biological relevance","volume":"20","author":"Tipper","year":"2008","journal-title":"J. Cognitive Neurosci."},{"key":"B34","doi-asserted-by":"publisher","first-page":"314","DOI":"10.1016\/j.jhevol.2006.10.001","article-title":"Reliance on head versus eyes in the gaze following of great apes and human infants: The cooperative eye hypothesis","volume":"52","author":"Tomasello","year":"2007","journal-title":"J. Hum. Evol."},{"key":"B35","first-page":"477","article-title":"Embodied social robots trigger gaze following in real-time HRI","author":"Wiese","year":"2018"}],"container-title":["Frontiers in Robotics and AI"],"original-title":[],"link":[{"URL":"https:\/\/www.frontiersin.org\/articles\/10.3389\/frobt.2023.1178433\/full","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,7,28]],"date-time":"2023-07-28T11:20:35Z","timestamp":1690543235000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.frontiersin.org\/articles\/10.3389\/frobt.2023.1178433\/full"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,7,28]]},"references-count":34,"alternative-id":["10.3389\/frobt.2023.1178433"],"URL":"https:\/\/doi.org\/10.3389\/frobt.2023.1178433","relation":{},"ISSN":["2296-9144"],"issn-type":[{"value":"2296-9144","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,7,28]]},"article-number":"1178433"}}