{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T04:16:18Z","timestamp":1775708178591,"version":"3.50.1"},"reference-count":43,"publisher":"Association for Computing Machinery (ACM)","issue":"ETRA","license":[{"start":{"date-parts":[[2024,5,20]],"date-time":"2024-05-20T00:00:00Z","timestamp":1716163200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["Proc. ACM Hum.-Comput. Interact."],"published-print":{"date-parts":[[2024,5,20]]},"abstract":"<jats:p>Successful learning depends on learners' ability to sustain attention, which is particularly challenging in online education due to limited teacher interaction. A potential indicator for attention is gaze synchrony, demonstrating predictive power for learning achievements in video-based learning in controlled experiments focusing on manipulating attention. This study (N=84) examines the relationship between gaze synchronization and self-reported attention of learners, using experience sampling, during realistic online video learning. Gaze synchrony was assessed through Kullback-Leibler Divergence of gaze density maps and MultiMatch algorithm scanpath comparisons. Results indicated significantly higher gaze synchronization in attentive participants for both measures and self-reported attention significantly predicted post-test scores. In contrast, synchrony measures did not correlate with learning outcomes. While supporting the hypothesis that attentive learners exhibit similar eye movements, the direct use of synchrony as an attention indicator poses challenges, requiring further research on the interplay of attention, gaze synchrony, and video content type.<\/jats:p>","DOI":"10.1145\/3655604","type":"journal-article","created":{"date-parts":[[2024,5,28]],"date-time":"2024-05-28T16:00:05Z","timestamp":1716912005000},"page":"1-18","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":12,"title":["On Task and in Sync: Examining the Relationship between Gaze Synchrony and Self-reported Attention During Video Lecture Learning"],"prefix":"10.1145","volume":"8","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1679-4979","authenticated-orcid":false,"given":"Babette","family":"B\u00fchler","sequence":"first","affiliation":[{"name":"Hector Research Institute of Education Sciences and Psychology, University of T\u00fcbingen, T\u00fcbingen, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4594-4318","authenticated-orcid":false,"given":"Efe","family":"Bozkir","sequence":"additional","affiliation":[{"name":"University of T\u00fcbingen &amp; Technical University of Munich, T\u00fcbingen, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8754-6123","authenticated-orcid":false,"given":"Hannah","family":"Deininger","sequence":"additional","affiliation":[{"name":"Hector Research Institute of Education Sciences and Psychology, University of T\u00fcbingen, T\u00fcbingen, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1358-6779","authenticated-orcid":false,"given":"Peter","family":"Gerjets","sequence":"additional","affiliation":[{"name":"Leibniz-Institut f\u00fcr Wissensmedien, T\u00fcbingen, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0647-0057","authenticated-orcid":false,"given":"Ulrich","family":"Trautwein","sequence":"additional","affiliation":[{"name":"Hector Research Institute of Education Sciences and Psychology, University of T\u00fcbingen, T\u00fcbingen, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3146-4484","authenticated-orcid":false,"given":"Enkelejda","family":"Kasneci","sequence":"additional","affiliation":[{"name":"Technical University of Munich, Munich, Germany"}]}],"member":"320","published-online":{"date-parts":[[2024,5,28]]},"reference":[{"key":"e_1_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0255979"},{"key":"e_1_2_1_2_1","volume-title":"Exploiting object-of-interest information to understand attention in VR classrooms. In 2021 IEEE Virtual Reality and 3D User Interfaces (VR)","author":"Bozkir Efe","unstructured":"Efe Bozkir, Philipp Stark, Hong Gao, Lisa Hasenbein, Jens-Uwe Hahn, Enkelejda Kasneci, and Richard G\u00f6llner. 2021b. Exploiting object-of-interest information to understand attention in VR classrooms. In 2021 IEEE Virtual Reality and 3D User Interfaces (VR). IEEE, 597--605."},{"key":"e_1_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijpsycho.2020.05.010"},{"key":"e_1_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0223941"},{"key":"e_1_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/3204493.3204550"},{"key":"e_1_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2021.3067787"},{"key":"e_1_2_1_7_1","volume-title":"It depends on how you look at it: Scanpath comparison in multiple dimensions with MultiMatch, a vector-based approach. Behavior research methods","author":"Dewhurst Richard","year":"2012","unstructured":"Richard Dewhurst, Marcus Nystr\u00f6m, Halszka Jarodzka, Tom Foulsham, Roger Johansson, and Kenneth Holmqvist. 2012. It depends on how you look at it: Scanpath comparison in multiple dimensions with MultiMatch, a vector-based approach. Behavior research methods , Vol. 44 (2012), 1079--1100."},{"key":"e_1_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1167\/10.10.28"},{"key":"e_1_2_1_9_1","volume-title":"Everyday attention and lecture retention: the effects of time, fidgeting, and mind wandering. Frontiers in psychology","author":"Farley James","year":"2013","unstructured":"James Farley, Evan F Risko, and Alan Kingstone. 2013. Everyday attention and lecture retention: the effects of time, fidgeting, and mind wandering. Frontiers in psychology , Vol. 4 (2013), 619."},{"key":"e_1_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.16910\/jemr.5.4.3"},{"key":"e_1_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445596"},{"key":"e_1_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10648-019-09514-z"},{"key":"e_1_2_1_13_1","volume-title":"Where people look when watching movies: Do all viewers look at the same place? Computers in biology and medicine","author":"Goldstein Robert B","year":"2007","unstructured":"Robert B Goldstein, Russell L Woods, and Eli Peli. 2007. Where people look when watching movies: Do all viewers look at the same place? Computers in biology and medicine, Vol. 37, 7 (2007), 957--964."},{"key":"e_1_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/1743666.1743718"},{"key":"e_1_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.learninstruc.2012.11.004"},{"key":"e_1_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1037\/xge0000362"},{"key":"e_1_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1111\/cogs.13247"},{"key":"e_1_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1007\/978--3-030--42504--3_15"},{"key":"e_1_2_1_19_1","volume-title":"Methods for comparing scanpaths and saliency maps: strengths and weaknesses. Behavior research methods","author":"Meur Olivier Le","year":"2013","unstructured":"Olivier Le Meur and Thierry Baccino. 2013. Methods for comparing scanpaths and saliency maps: strengths and weaknesses. Behavior research methods, Vol. 45, 1 (2013), 251--266."},{"key":"e_1_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/2638728.2641688"},{"key":"e_1_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.lindif.2010.12.006"},{"key":"e_1_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1007\/s00426-023-01791--2"},{"key":"e_1_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.2016980118"},{"key":"e_1_2_1_24_1","volume-title":"Human vision and electronic imaging IX","author":"Rajashekar Umesh","unstructured":"Umesh Rajashekar, Lawrence K Cormack, and Alan C Bovik. 2004. Point-of-gaze analysis reveals visual search strategies. In Human vision and electronic imaging IX, Vol. 5292. SPIE, 296--306."},{"key":"e_1_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3517031.3529239"},{"key":"e_1_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3603555.3603560"},{"key":"e_1_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3517031.3529620"},{"key":"e_1_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2007.10.010"},{"key":"e_1_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1037\/stl0000011"},{"key":"e_1_2_1_30_1","volume-title":"An Integrated Approach","author":"Schiffman H.R.","unstructured":"H.R. Schiffman. 2001. Sensation and Perception. An Integrated Approach. John Wiley and Sons."},{"key":"e_1_2_1_31_1","unstructured":"Sensomotoric Instruments. 2017. BeGaze Manual Version 3.7."},{"key":"e_1_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1167\/13.8.16"},{"key":"e_1_2_1_33_1","doi-asserted-by":"publisher","unstructured":"\u00d6mer S\u00fcmer Peter Gerjets Ulrich Trautwein and Enkelejda Kasneci. 2020. Automated Anonymisation of Visual and Audio Data in Classroom Studies. https:\/\/doi.org\/10.48550\/arXiv.2001.05080 The Workshops of the Thirty-Fourth AAAI Conference on Artificial Intelligence.","DOI":"10.48550\/arXiv.2001.05080"},{"key":"e_1_2_1_34_1","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops. 2315--2324","author":"S\u00fcmer \u00d6mer","year":"2018","unstructured":"\u00d6mer S\u00fcmer, Patricia Goldberg, Kathleen Sturmer, Tina Seidel, Peter Gerjets, Ulrich Trautwein, and Enkelejda Kasneci. 2018. Teachers' perception in the classroom. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops. 2315--2324."},{"key":"e_1_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3359264"},{"key":"e_1_2_1_36_1","volume-title":"Visual correlates of fixation selection: Effects of scale and time. Vision research","author":"Tatler Benjamin W","year":"2005","unstructured":"Benjamin W Tatler, Roland J Baddeley, and Iain D Gilchrist. 2005. Visual correlates of fixation selection: Effects of scale and time. Vision research, Vol. 45, 5 (2005), 643--659."},{"key":"e_1_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.21105\/joss.01525"},{"key":"e_1_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.compedu.2018.12.007"},{"key":"e_1_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1037\/h0101808"},{"key":"e_1_2_1_40_1","volume-title":"how do I measure thee with probes? Let me count the ways. Behavior research methods","author":"Weinstein Yana","year":"2018","unstructured":"Yana Weinstein. 2018. Mind-wandering, how do I measure thee with probes? Let me count the ways. Behavior research methods , Vol. 50 (2018), 642--661."},{"key":"e_1_2_1_41_1","volume-title":"EdMedia Innovate Learning","author":"Wisiecka Katarzyna","unstructured":"Katarzyna Wisiecka, Krzysztof Krejtz, Izabela Krejtz, and Andrew Duchowski. 2022. Dynamics of visual attention during online lectures-evidence from webcam eye tracking. In EdMedia Innovate Learning. Association for the Advancement of Computing in Education (AACE), 1220--1230."},{"key":"e_1_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.3758\/BF03195480"},{"key":"e_1_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1002\/acp.3632"}],"container-title":["Proceedings of the ACM on Human-Computer Interaction"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3655604","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3655604","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,23]],"date-time":"2025-08-23T01:23:01Z","timestamp":1755912181000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3655604"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,20]]},"references-count":43,"journal-issue":{"issue":"ETRA","published-print":{"date-parts":[[2024,5,20]]}},"alternative-id":["10.1145\/3655604"],"URL":"https:\/\/doi.org\/10.1145\/3655604","relation":{},"ISSN":["2573-0142"],"issn-type":[{"value":"2573-0142","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,5,20]]},"assertion":[{"value":"2024-05-28","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}