{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T19:11:03Z","timestamp":1776107463974,"version":"3.50.1"},"reference-count":51,"publisher":"Frontiers Media SA","license":[{"start":{"date-parts":[[2023,4,10]],"date-time":"2023-04-10T00:00:00Z","timestamp":1681084800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100002428","name":"Austrian Science Fund","doi-asserted-by":"publisher","award":["I3969-N30"],"award-info":[{"award-number":["I3969-N30"]}],"id":[{"id":"10.13039\/501100002428","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["frontiersin.org"],"crossmark-restriction":true},"short-container-title":["Front. Robot. AI"],"abstract":"<jats:p>Similar to human\u2013human interaction (HHI), gaze is an important modality in conversational human\u2013robot interaction (HRI) settings. Previously, human-inspired gaze parameters have been used to implement gaze behavior for humanoid robots in conversational settings and improve user experience (UX). Other robotic gaze implementations disregard social aspects of gaze behavior and pursue a technical goal (e.g., face tracking). However, it is unclear how deviating from human-inspired gaze parameters affects the UX. In this study, we use eye-tracking, interaction duration, and self-reported attitudinal measures to study the impact of non-human inspired gaze timings on the UX of the participants in a conversational setting. We show the results for systematically varying the gaze aversion ratio (GAR) of a humanoid robot over a broad parameter range from almost always gazing at the human conversation partner to almost always averting the gaze. The main results reveal that on a behavioral level, a low GAR leads to shorter interaction durations and that human participants change their GAR to mimic the robot. However, they do not copy the robotic gaze behavior strictly. Additionally, in the lowest gaze aversion setting, participants do not gaze back as much as expected, which indicates a user aversion to the robot gaze behavior. However, participants do not report different attitudes toward the robot for different GARs during the interaction. In summary, the urge of humans in conversational settings with a humanoid robot to adapt to the perceived GAR is stronger than the urge of intimacy regulation through gaze aversion, and a high mutual gaze is not always a sign\u00a0of high comfort, as suggested earlier. This result can be used as a justification to deviate from human-inspired gaze parameters when necessary for specific robot behavior implementations.<\/jats:p>","DOI":"10.3389\/frobt.2023.1062714","type":"journal-article","created":{"date-parts":[[2023,4,10]],"date-time":"2023-04-10T04:20:08Z","timestamp":1681100408000},"update-policy":"https:\/\/doi.org\/10.3389\/crossmark-policy","source":"Crossref","is-referenced-by-count":5,"title":["Robotic gaze and human views: A systematic exploration of robotic gaze aversion and its effects on human behaviors and attitudes"],"prefix":"10.3389","volume":"10","author":[{"given":"Michael","family":"Koller","sequence":"first","affiliation":[]},{"given":"Astrid","family":"Weiss","sequence":"additional","affiliation":[]},{"given":"Matthias","family":"Hirschmanner","sequence":"additional","affiliation":[]},{"given":"Markus","family":"Vincze","sequence":"additional","affiliation":[]}],"member":"1965","published-online":{"date-parts":[[2023,4,10]]},"reference":[{"key":"B1","doi-asserted-by":"publisher","DOI":"10.16910\/jemr.14.1.1","article-title":"Gaze aversion in conversational settings: An investigation based on mock job interview","volume":"14","author":"Acarturk","year":"2021","journal-title":"J. Eye Mov. Res."},{"key":"B2","first-page":"1983","article-title":"Robot gaze does not reflexively cue human attention","volume":"33","author":"Admoni","year":"2011","journal-title":"Annu. Meet. Cognitive Sci. Soc."},{"key":"B3","doi-asserted-by":"publisher","first-page":"25","DOI":"10.5898\/jhri.6.1.admoni","article-title":"Social eye gaze in human-robot interaction: A review","volume":"6","author":"Admoni","year":"2017","journal-title":"J. Human-Robot Interact."},{"key":"B4","first-page":"25","article-title":"Conversational gaze aversion for humanlike robots","volume-title":"2014 9th ACM\/IEEE international conference on human-robot interaction","author":"Andrist","year":"2014"},{"key":"B5","doi-asserted-by":"publisher","first-page":"848","DOI":"10.1017\/s0007125000073980","article-title":"Gaze and mutual gaze","volume":"165","author":"Argyle","year":"1994","journal-title":"Br. J. Psychiatry"},{"key":"B6","doi-asserted-by":"publisher","first-page":"289","DOI":"10.2307\/2786027","article-title":"Eye-contact, distance and affiliation","volume":"28","author":"Argyle","year":"1965","journal-title":"Sociometry"},{"key":"B7","volume-title":"The psychology of interpersonal behaviour","author":"Argyle","year":"1994"},{"key":"B8","first-page":"4144","article-title":"Tracking a varying number of people with a visually-controlled robotic head","volume-title":"IEEE\/RSJ international conference on intelligent robots and systems","author":"Ban","year":"2017"},{"key":"B9","doi-asserted-by":"publisher","first-page":"71","DOI":"10.1007\/s12369-008-0001-3","article-title":"Measurement instruments for the anthropomorphism, animacy, likeability, perceived intelligence, and perceived safety of robots","volume":"1","author":"Bartneck","year":"2009","journal-title":"Int. J. Soc. Robotics"},{"key":"B10","first-page":"126","article-title":"Tracking gaze over time in hri as a proxy for engagement and attribution of social agency","volume-title":"2014 9th ACM\/IEEE international conference on human-robot interaction","author":"Baxter","year":"2014"},{"key":"B11","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1080\/00220973.1995.9943797","article-title":"Multiple regression approach to analyzing contingency tables: Post hoc and planned comparison procedures","volume":"64","author":"Beasley","year":"1995","journal-title":"J. Exp. Educ."},{"key":"B12","doi-asserted-by":"publisher","first-page":"276","DOI":"10.1037\/0033-2909.91.2.276","article-title":"Task-evoked pupillary responses, processing load, and the structure of processing resources","volume":"91","author":"Beatty","year":"1982","journal-title":"Psychol. Bull."},{"key":"B13","doi-asserted-by":"crossref","first-page":"41","DOI":"10.1007\/978-1-4471-4784-8_4","article-title":"Automated analysis of mutual gaze in human conversational pairs","volume-title":"Eye gaze in intelligent user interfaces","author":"Broz","year":"2013"},{"key":"B14","first-page":"858","article-title":"Mutual gaze, personality, and familiarity: Dual eye-tracking during conversation","volume-title":"2012 21st IEEE international symposium on robot and human interactive communication (RO-MAN)","author":"Broz","year":"2012"},{"key":"B15","doi-asserted-by":"publisher","first-page":"104734","DOI":"10.1016\/j.cognition.2021.104734","article-title":"Meeting another\u2019s gaze shortens subjective time by capturing attention","volume":"212","author":"Burra","year":"2021","journal-title":"Cognition"},{"key":"B16","doi-asserted-by":"publisher","first-page":"89","DOI":"10.1007\/s12193-009-0033-5","article-title":"Affect recognition for interactive companions: Challenges and design in real world scenarios","volume":"3","author":"Castellano","year":"2010","journal-title":"J. Multimodal User Interfaces"},{"key":"B17","doi-asserted-by":"publisher","first-page":"59","DOI":"10.1007\/s11920-017-0808-4","article-title":"Gaze-based assessments of vigilance and avoidance in social anxiety: A review","volume":"19","author":"Chen","year":"2017","journal-title":"Curr. Psychiatry Rep."},{"key":"B18","doi-asserted-by":"crossref","DOI":"10.4324\/9780203771587","volume-title":"Statistical power analysis for the behavioral sciences","author":"Cohen","year":"2013"},{"key":"B19","doi-asserted-by":"publisher","first-page":"581","DOI":"10.1016\/s0149-7634(00)00025-7","article-title":"The eyes have it: The neuroethology, function and evolution of social gaze","volume":"24","author":"Emery","year":"2000","journal-title":"Neurosci. Biobehav. Rev."},{"key":"B20","doi-asserted-by":"publisher","first-page":"1149","DOI":"10.3758\/brm.41.4.1149","article-title":"Statistical power analyses using g* power 3.1: Tests for correlation and regression analyses","volume":"41","author":"Faul","year":"2009","journal-title":"Behav. Res. Methods"},{"key":"B21","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1016\/0885-2308(91)90019-m","article-title":"Simulating speech systems","volume":"5","author":"Fraser","year":"1991","journal-title":"Comput. Speech Lang."},{"key":"B22","doi-asserted-by":"publisher","first-page":"825","DOI":"10.1177\/0013164403251280","article-title":"Cellwise residual analysis in two-way contingency tables","volume":"63","author":"Garcia-Perez","year":"2003","journal-title":"Educ. Psychol. Meas."},{"key":"B23","doi-asserted-by":"publisher","first-page":"107713","DOI":"10.1016\/j.biopsycho.2019.06.001","article-title":"Eye gaze behavior during affective picture viewing: Effects of motivational significance, gender, age, and repeated exposure","volume":"146","author":"Gomez","year":"2019","journal-title":"Biol. Psychol."},{"key":"B24","first-page":"175","article-title":"Investigating transparency methods in a robot word-learning system and their effects on human teaching behaviors","volume-title":"2021 30th IEEE international conference on robot and human interactive communication (RO-MAN)","author":"Hirschmanner","year":"2021"},{"key":"B25","first-page":"504","article-title":"Exploring gaze behaviour and perceived personality traits","volume-title":"2020 15th ACM\/IEEE international conference on human-computer interaction","author":"Ijuin","year":"2020"},{"key":"B26","volume-title":"The illusion of life: Disney animation (Disney Editions New York)","author":"Johnston","year":"1981"},{"key":"B27","first-page":"1","article-title":"Conversational gaze modelling in first encounter robot dialogues","volume-title":"11th international conference on language resources and evaluation","author":"Jokinen","year":"2018"},{"key":"B28","doi-asserted-by":"publisher","first-page":"379","DOI":"10.1007\/s12369-013-0191-1","article-title":"Behave-ii: The revised set of measures to assess users\u2019 attitudinal and behavioral responses to a social robot","volume":"5","author":"Joosse","year":"2013","journal-title":"Int. J. Soc. Robotics"},{"key":"B29","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1016\/j.neuron.2015.11.028","article-title":"Relationships between pupil diameter and neuronal activity in the locus coeruleus, colliculi, and cingulate cortex","volume":"89","author":"Joshi","year":"2016","journal-title":"Neuron"},{"key":"B30","first-page":"983","article-title":"Averting robot eyes","volume":"76","author":"Kaminski","year":"2016","journal-title":"Md. L. Rev."},{"key":"B31","doi-asserted-by":"publisher","first-page":"22","DOI":"10.1016\/0001-6918(67)90005-4","article-title":"Some functions of gaze-direction in social interaction","volume":"26","author":"Kendon","year":"1967","journal-title":"Acta Psychol."},{"key":"B32","doi-asserted-by":"publisher","first-page":"293","DOI":"10.1007\/s12369-014-0277-4","article-title":"Comparing robot embodiments in a guided discovery learning interaction with children","volume":"7","author":"Kennedy","year":"2015","journal-title":"Int. J. Soc. Robotics"},{"key":"B33","first-page":"644","article-title":"A pilot study on determining the relation between gaze aversion and interaction experience","volume-title":"2019 14th ACM\/IEEE international conference on human-robot interaction","author":"Koller","year":"2019"},{"key":"B34","doi-asserted-by":"crossref","first-page":"526","DOI":"10.1007\/978-3-319-70022-9_52","article-title":"Naturalistic conversational gaze control for humanoid robots-a first step","volume-title":"International conference on social robotics","author":"Lehmann","year":"2017"},{"key":"B35","doi-asserted-by":"crossref","first-page":"519","DOI":"10.1109\/ROMAN.2016.7745167","article-title":"Playing the \u2018trust game\u2019with robots: Social strategies and experiences","volume-title":"2016 25th IEEE international symposium on robot and human interactive communication (RO-MAN)","author":"Mota","year":"2016"},{"key":"B36","doi-asserted-by":"publisher","first-page":"321","DOI":"10.1075\/is.17.3.01pap","article-title":"Do relative positions and proxemics affect the engagement in a human-robot collaborative scenario?","volume":"17","author":"Papadopoulos","year":"2016","journal-title":"Interact. Stud."},{"key":"B37","first-page":"515","article-title":"Brownian movement and markov processes","volume-title":"Probability, random variables, and stochastic processes","author":"Papoulis","year":"1984"},{"key":"B38","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1016\/s1071-5819(03)00017-x","article-title":"Pupil size variation as an indication of affective processing","volume":"59","author":"Partala","year":"2003","journal-title":"Int. J. Human-Computer Stud."},{"key":"B39","doi-asserted-by":"publisher","first-page":"293","DOI":"10.1098\/rspb.1985.0003","article-title":"Visual cells in the temporal cortex sensitive to face view and gaze direction","volume":"223","author":"Perrett","year":"1985","journal-title":"Proc. R. Soc. Lond. Ser. B. Biol. Sci."},{"key":"B40","doi-asserted-by":"publisher","first-page":"645956","DOI":"10.3389\/frobt.2021.645956","article-title":"I can see it in your eyes: Gaze as an implicit cue of uncanniness and task performance in repeated interactions with robots","volume":"8","author":"Perugia","year":"2021","journal-title":"Front. Robotics AI"},{"key":"B41","doi-asserted-by":"publisher","first-page":"275911","DOI":"10.1371\/journal.pone.0027591","article-title":"A non-verbal turing test: Differentiating mind from machine in gaze-based social interaction","volume":"6","author":"Pfeiffer","year":"2011","journal-title":"PLOS ONE"},{"key":"B42","first-page":"19","volume-title":"The media equation: How people treat computers, television, and new media like real people","author":"Reeves","year":"1996"},{"key":"B43","volume-title":"Nutty-based robot animation\u2013principles and practices","author":"Ribeiro","year":"2019"},{"key":"B44","doi-asserted-by":"publisher","first-page":"119","DOI":"10.5898\/jhri.1.1.riek","article-title":"Wizard of oz studies in hri: A systematic review and new reporting guidelines","volume":"1","author":"Riek","year":"2012","journal-title":"J. Human-Robot Interact."},{"key":"B45","first-page":"1","article-title":"Gaze and attention during an hri storytelling task","volume-title":"2015 AAAI fall symposium series","author":"Sabyruly","year":"2015"},{"key":"B46","doi-asserted-by":"publisher","first-page":"140","DOI":"10.1016\/j.artint.2005.03.005","article-title":"Explorations in engagement for humans and robots","volume":"166","author":"Sidner","year":"2005","journal-title":"Artif. Intell."},{"key":"B47","first-page":"253","article-title":"A survey of social gaze","volume-title":"2011 6th ACM\/IEEE international conference on human-robot interaction","author":"Srinivasan","year":"2011"},{"key":"B48","first-page":"82","article-title":"The psychology of rapport: Five basic rules","author":"St-Yves","year":"2006","journal-title":"Investig. Interviewing"},{"key":"B49","first-page":"11","article-title":"The usus evaluation framework for human-robot interaction","volume-title":"AISB2009: Symposium on new Frontiers in human-robot interaction","author":"Weiss","year":"2009"},{"key":"B50","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2070719.2070726","article-title":"Adaptive eye gaze patterns in interactions with human and artificial agents","volume":"1","author":"Yu","year":"2012","journal-title":"ACM Trans. Interact. Intelligent Syst."},{"key":"B51","doi-asserted-by":"crossref","first-page":"556","DOI":"10.1007\/978-3-319-70022-9_55","article-title":"Look but don\u2019t stare: Mutual gaze interaction in social robots","volume-title":"International conference on social robotics","author":"Zhang","year":"2017"}],"container-title":["Frontiers in Robotics and AI"],"original-title":[],"link":[{"URL":"https:\/\/www.frontiersin.org\/articles\/10.3389\/frobt.2023.1062714\/full","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,4,10]],"date-time":"2023-04-10T04:20:21Z","timestamp":1681100421000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.frontiersin.org\/articles\/10.3389\/frobt.2023.1062714\/full"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4,10]]},"references-count":51,"alternative-id":["10.3389\/frobt.2023.1062714"],"URL":"https:\/\/doi.org\/10.3389\/frobt.2023.1062714","relation":{},"ISSN":["2296-9144"],"issn-type":[{"value":"2296-9144","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,4,10]]},"article-number":"1062714"}}