{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,13]],"date-time":"2026-01-13T04:19:42Z","timestamp":1768277982876,"version":"3.49.0"},"reference-count":43,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"12","license":[{"start":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T00:00:00Z","timestamp":1733011200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"DOI":"10.13039\/100004318","name":"Microsoft","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100004318","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100008473","name":"Bournemouth University","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100008473","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Pattern Anal. Mach. Intell."],"published-print":{"date-parts":[[2024,12]]},"DOI":"10.1109\/tpami.2024.3411938","type":"journal-article","created":{"date-parts":[[2024,6,10]],"date-time":"2024-06-10T17:24:19Z","timestamp":1718040259000},"page":"8715-8726","source":"Crossref","is-referenced-by-count":4,"title":["Dual Input Stream Transformer for Vertical Drift Correction in Eye-Tracking Reading Data"],"prefix":"10.1109","volume":"46","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8367-0346","authenticated-orcid":false,"given":"Thomas M.","family":"Mercier","sequence":"first","affiliation":[{"name":"Faculty of Science &amp; Technology, Bournemouth University, Poole, U.K."}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0158-6309","authenticated-orcid":false,"given":"Marcin","family":"Budka","sequence":"additional","affiliation":[{"name":"Faculty of Science &amp; Technology, Bournemouth University, Poole, U.K."}]},{"given":"Martin R.","family":"Vasilev","sequence":"additional","affiliation":[{"name":"Faculty of Brain Sciences, University College London, London, U.K."}]},{"given":"Julie A.","family":"Kirkby","sequence":"additional","affiliation":[{"name":"Faculty of Science &amp; Technology, Bournemouth University, Poole, U.K."}]},{"given":"Bernhard","family":"Angele","sequence":"additional","affiliation":[{"name":"Nebrija Research Centre in Cognition (Centro de Investigaci&#x00F3;n Nebrija en Cognici&#x00F3;n, CINC), Universidad Antonio de Nebrija, Madrid, Spain"}]},{"given":"Timothy J.","family":"Slattery","sequence":"additional","affiliation":[{"name":"Faculty of Science &amp; Technology, Bournemouth University, Poole, U.K."}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-20085-5_8"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1016\/j.array.2021.100087"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1080\/17470210902816461"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ZINC55034.2022.9840532"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1093\/schbul\/sbaa107"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1080\/0144929X.2018.1551933"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.3758\/s13428-021-01554-0"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1080\/10888438.2023.2259522"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1002\/acp.3195"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1016\/j.visres.2021.01.003"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.3758\/s13428-017-0955-x"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.3758\/s13428-012-0280-3"},{"key":"ref13","article-title":"Algorithms for assigning fixations to lines of text in multiline passage reading","volume-title":"21st Eur. Conf. Eye Movements","author":"Carr","year":"2022"},{"key":"ref14","article-title":"popEye - An R package to analyse eye movement data from reading experiments","author":"Schroeder","year":"2019","journal-title":"GitHub repository"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/TASSP.1978.1163055"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1007\/BF01074755"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.3758\/s13428-018-1120-x"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1016\/j.procs.2021.09.069"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1145\/2745555.2746644"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.23919\/EUSIPCO55093.2022.9909817"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/n18-1180"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1145\/3379156.3391335"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1145\/3517031.3529639"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1145\/3591131"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1007\/s00426-021-01610-6"},{"key":"ref26","article-title":"Measured and perceived impact of noise during reading. an eye tracking study","author":"Goldenberg","year":"2022"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1037\/xhp0000680"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.3758\/s13414-019-01742-3"},{"key":"ref29","article-title":"Using eye-tracking technology to measure the effect of political ideology on comprehension of online news articles","author":"Seymour","year":"2022"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.3758\/s13428-021-01772-6"},{"key":"ref31","article-title":"Pointer sentinel mixture models","author":"Merity","year":"2016"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1810.04805"},{"key":"ref33","article-title":"Attention is all you need","author":"Vaswani","year":"2017"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/ASRU46091.2019.9004025"},{"key":"ref35","volume-title":"Deep Learning, Ser. Adaptive Computation and Machine Learning","author":"Goodfellow","year":"2016"},{"key":"ref36","first-page":"3965","article-title":"CoAtNet: Marrying convolution and attention for all data sizes","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Dai","year":"2021"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref38","article-title":"PyTorch image models","author":"Wightman","year":"2019","journal-title":"GitHub repository"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1007\/s10044-023-01181-9"},{"key":"ref40","article-title":"PyTorch: An imperative style, high-performance deep learning library","author":"Paszke","year":"2019"},{"key":"ref41","article-title":"Well-read students learn better: On the importance of pre-training compact models","author":"Turc","year":"2019"},{"key":"ref42","article-title":"Eyekit","year":"2019"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.123"}],"container-title":["IEEE Transactions on Pattern Analysis and Machine Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/34\/10746266\/10552407.pdf?arnumber=10552407","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,27]],"date-time":"2024-11-27T00:25:46Z","timestamp":1732667146000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10552407\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12]]},"references-count":43,"journal-issue":{"issue":"12"},"URL":"https:\/\/doi.org\/10.1109\/tpami.2024.3411938","relation":{},"ISSN":["0162-8828","2160-9292","1939-3539"],"issn-type":[{"value":"0162-8828","type":"print"},{"value":"2160-9292","type":"electronic"},{"value":"1939-3539","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12]]}}}