{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T05:05:34Z","timestamp":1750309534884,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":81,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,5,25]],"date-time":"2025-05-25T00:00:00Z","timestamp":1748131200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-sa\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,5,26]]},"DOI":"10.1145\/3715669.3723118","type":"proceedings-article","created":{"date-parts":[[2025,5,24]],"date-time":"2025-05-24T06:57:59Z","timestamp":1748069879000},"page":"1-8","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Exploring promptable foundation models for high-resolution video eye tracking in the lab"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4672-8756","authenticated-orcid":false,"given":"Diederick C.","family":"Niehorster","sequence":"first","affiliation":[{"name":"Lund University Humanities Lab and Department of Psychology, Lund University, Lund, Sweden"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-5580-2304","authenticated-orcid":false,"given":"Virmarie","family":"Maquiling","sequence":"additional","affiliation":[{"name":"Human-Centered Technologies for Learning, Technical University of Munich, Munich, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-5685-7318","authenticated-orcid":false,"given":"Sean","family":"Byrne","sequence":"additional","affiliation":[{"name":"Dipartimento di Elettronica, Informazione e Bioingegneria, Politecnico di Milano, Milan, Italy"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3146-4484","authenticated-orcid":false,"given":"Enkelejda","family":"Kasneci","sequence":"additional","affiliation":[{"name":"Human-Centered Technologies for Learning, Technical University of Munich, Munich, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2089-9012","authenticated-orcid":false,"given":"Marcus","family":"Nystr\u00f6m","sequence":"additional","affiliation":[{"name":"Lund University Humanities Lab, Lund, Sweden"}]}],"member":"320","published-online":{"date-parts":[[2025,5,25]]},"reference":[{"key":"e_1_3_3_1_2_1","unstructured":"Josh Achiam Steven Adler Sandhini Agarwal Lama Ahmad Ilge Akkaya Florencia\u00a0Leoni Aleman Diogo Almeida Janko Altenschmidt Sam Altman Shyamal Anadkat et\u00a0al. 2023. Gpt-4 technical report. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2303.08774 (2023)."},{"key":"e_1_3_3_1_3_1","doi-asserted-by":"crossref","unstructured":"Andronicus\u00a0A Akinyelu and Pieter Blignaut. 2020. Convolutional neural network-based methods for eye gaze estimation: A survey. IEEE Access 8 (2020) 142581\u2013142605.","DOI":"10.1109\/ACCESS.2020.3013540"},{"key":"e_1_3_3_1_4_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2006.312773"},{"key":"e_1_3_3_1_5_1","unstructured":"Sean\u00a0Anthony Byrne Virmarie Maquiling Marcus Nystr\u00f6m Enkelejda Kasneci and Diederick\u00a0C. Niehorster. 2023. LEyes: A Lightweight Framework for Deep Learning-Based Eye Tracking using Synthetic Eye Images. arXiv abs\/2309.06129 (2023). arXiv:https:\/\/arXiv.org\/abs\/2309.06129https:\/\/arxiv.org\/abs\/2309.06129"},{"key":"e_1_3_3_1_6_1","doi-asserted-by":"publisher","unstructured":"Sean\u00a0Anthony Byrne Marcus Nystr\u00f6m Virmarie Maquiling Enkelejda Kasneci and Diederick\u00a0C. Niehorster. 2024. Precise localization of corneal reflections in eye images using deep learning trained on synthetic data. Behavior Research Methods 56 4 (01 6 2024) 3226\u20133241. 10.3758\/s13428-023-02297-w","DOI":"10.3758\/s13428-023-02297-w"},{"key":"e_1_3_3_1_7_1","unstructured":"Jiazhong Cen Zanwei Zhou Jiemin Fang Wei Shen Lingxi Xie Dongsheng Jiang Xiaopeng Zhang Qi Tian et\u00a0al. 2023. Segment anything in 3d with nerfs. Advances in Neural Information Processing Systems 36 (2023) 25971\u201325990."},{"key":"e_1_3_3_1_8_1","doi-asserted-by":"publisher","unstructured":"Juan\u00a0J Cerrolaza Arantxa Villanueva and Rafael Cabeza. 2012. Study of polynomial mapping functions in video-oculography eye trackers. ACM Transactions on Computer-Human Interaction 19 2 (2012) 1\u201325. 10.1145\/2240156.2240158","DOI":"10.1145\/2240156.2240158"},{"key":"e_1_3_3_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2019.00568"},{"key":"e_1_3_3_1_10_1","doi-asserted-by":"crossref","unstructured":"Ho\u00a0Kei Cheng Seoung\u00a0Wug Oh Brian Price Alexander Schwing and Joon-Young Lee. 2023b. Tracking Anything with Decoupled Video Segmentation. arXiv:https:\/\/arXiv.org\/abs\/2309.03903https:\/\/arxiv.org\/abs\/2309.03903","DOI":"10.1109\/ICCV51070.2023.00127"},{"key":"e_1_3_3_1_11_1","unstructured":"Yangming Cheng Liulei Li Yuanyou Xu Xiaodi Li Zongxin Yang Wenguan Wang and Yi Yang. 2023a. Segment and Track Anything. arXiv:https:\/\/arXiv.org\/abs\/2305.06558https:\/\/arxiv.org\/abs\/2305.06558"},{"key":"e_1_3_3_1_12_1","unstructured":"Yihua Cheng Haofei Wang Yiwei Bao and Feng Lu. 2024. Appearance-based Gaze Estimation With Deep Learning: A Review and Benchmark. arXiv abs\/2104.12668 (2024). arXiv:https:\/\/arXiv.org\/abs\/2104.12668http:\/\/arxiv.org\/abs\/2104.12668"},{"key":"e_1_3_3_1_13_1","unstructured":"Jiangfan Deng Zhuang Jia Zhaoxue Wang Xiang Long and Daniel\u00a0K. Du. 2024. Towards Unsupervised Eye-Region Segmentation for Eye Tracking. arXiv:https:\/\/arXiv.org\/abs\/2410.06131https:\/\/arxiv.org\/abs\/2410.06131"},{"key":"e_1_3_3_1_14_1","doi-asserted-by":"crossref","unstructured":"Lei Ding Kun Zhu Daifeng Peng Hao Tang Kuiwu Yang and Lorenzo Bruzzone. 2024b. Adapting segment anything model for change detection in VHR remote sensing images. IEEE Transactions on Geoscience and Remote Sensing (2024).","DOI":"10.1109\/TGRS.2024.3368168"},{"key":"e_1_3_3_1_15_1","unstructured":"Shuangrui Ding Rui Qian Xiaoyi Dong Pan Zhang Yuhang Zang Yuhang Cao Yuwei Guo Dahua Lin and Jiaqi Wang. 2024a. SAM2Long: Enhancing SAM 2 for Long Video Segmentation with a Training-Free Memory Tree. arXiv abs\/2410.16268 (2024). arXiv:https:\/\/arXiv.org\/abs\/2410.16268https:\/\/arxiv.org\/abs\/2410.16268"},{"key":"e_1_3_3_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3314111.3319914"},{"key":"e_1_3_3_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3379156.3391347"},{"key":"e_1_3_3_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3204493.3204559"},{"key":"e_1_3_3_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISMAR52148.2021.00053"},{"key":"e_1_3_3_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/2968219.2968334"},{"key":"e_1_3_3_1_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/WACV.2017.126"},{"key":"e_1_3_3_1_22_1","unstructured":"Wolfgang Fuhl Thiago Santini Gjergji Kasneci and Enkelejda Kasneci. 2016b. PupilNet: convolutional neural networks for robust pupil detection. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/1601.04902 (2016)."},{"key":"e_1_3_3_1_23_1","unstructured":"Wolfgang Fuhl Thiago Santini Gjergji Kasneci Wolfgang Rosenstiel and Enkelejda Kasneci. 2017b. PupilNet v2. 0: Convolutional Neural Networks for CPU based real time Robust Pupil Detection. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/1711.00112 (2017)."},{"key":"e_1_3_3_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/2857491.2857505"},{"key":"e_1_3_3_1_25_1","doi-asserted-by":"publisher","DOI":"10.5220\/0011607200003417"},{"key":"e_1_3_3_1_26_1","doi-asserted-by":"publisher","unstructured":"Wolfgang Fuhl Daniel Weber and Shahram Eivazi. 2024. Pistol: Pupil Invisible Supportive Tool in the Wild. SN Computer Science 5 3 (21 2 2024) 276. 10.1007\/s42979-024-02606-w","DOI":"10.1007\/s42979-024-02606-w"},{"key":"e_1_3_3_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3379155.3391317"},{"key":"e_1_3_3_1_28_1","unstructured":"Ziyu Guo Renrui Zhang Xiangyang Zhu Chengzhuo Tong Peng Gao Chunyuan Li and Pheng-Ann Heng. 2024. Sam2point: Segment any 3d as videos in zero-shot and promptable manners. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2408.16768 (2024)."},{"key":"e_1_3_3_1_29_1","doi-asserted-by":"publisher","unstructured":"Roy\u00a0S. Hessels Antje Nuthmann Marcus Nystr\u00f6m Richard Andersson Diederick\u00a0C. Niehorster and Ignace T.\u00a0C. Hooge. 2024. The fundamentals of eye tracking part 1: The link between theory and research question. Behavior Research Methods 57 1 (12 12 2024) 16. 10.3758\/s13428-024-02544-8","DOI":"10.3758\/s13428-024-02544-8"},{"key":"e_1_3_3_1_30_1","doi-asserted-by":"publisher","unstructured":"Ignace T.\u00a0C. Hooge Roy\u00a0S. Hessels Diederick\u00a0C. Niehorster Richard Andersson Marta\u00a0K. Skrok Robert Konklewski Patrycjusz Stremplewski Maciej Nowakowski Szymon Tamborski Anna Szkulmowska Maciej Szkulmowski and Marcus Nystr\u00f6m. 2024. Eye tracker calibration: How well can humans refixate a target?Behavior Research Methods 57 1 (2024) 23. 10.3758\/s13428-024-02564-4","DOI":"10.3758\/s13428-024-02564-4"},{"key":"e_1_3_3_1_31_1","doi-asserted-by":"publisher","unstructured":"Ignace T.\u00a0C. Hooge Diederick\u00a0C. Niehorster Roy\u00a0S. Hessels Jeroen\u00a0S. Benjamins and Marcus Nystr\u00f6m. 2022. How robust are wearable eye trackers to slow and fast head and body movements?Behavior Research Methods (2022) 1\u201315. 10.3758\/s13428-022-02010-3","DOI":"10.3758\/s13428-022-02010-3"},{"key":"e_1_3_3_1_32_1","doi-asserted-by":"publisher","unstructured":"Ignace T\u00a0C Hooge Diederick\u00a0C Niehorster Roy\u00a0S Hessels Dixon Cleveland and Marcus Nystr\u00f6m. 2021. The pupil-size artefact (PSA) across time viewing direction and different eye trackers. Behavior Research Methods (2021). 10.3758\/s13428-020-01512-2","DOI":"10.3758\/s13428-020-01512-2"},{"key":"e_1_3_3_1_33_1","doi-asserted-by":"publisher","unstructured":"Ignace T.\u00a0C. Hooge Diederick\u00a0C. Niehorster Marcus Nystr\u00f6m Richard Andersson and Roy\u00a0S. Hessels. 2018. Is human classification by experienced untrained observers a gold standard in fixation detection?Behavior Research Methods 50 5 (2018) 1864\u20131881. 10.3758\/s13428-017-0955-x","DOI":"10.3758\/s13428-017-0955-x"},{"key":"e_1_3_3_1_34_1","doi-asserted-by":"crossref","unstructured":"Yuhao Huang Xin Yang Lian Liu Han Zhou Ao Chang Xinrui Zhou Rusi Chen Junxuan Yu Jiongquan Chen Chaoyu Chen et\u00a0al. 2024. Segment anything model for medical images?Medical Image Analysis 92 (2024) 103061.","DOI":"10.1016\/j.media.2023.103061"},{"key":"e_1_3_3_1_35_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2019.00457"},{"key":"e_1_3_3_1_36_1","unstructured":"Zhuang Jia Jiangfan Deng Liying Chi Xiang Long and Daniel\u00a0K. Du. 2024. CondSeg: Ellipse Estimation of Pupil and Iris via Conditioned Segmentation. arXiv:https:\/\/arXiv.org\/abs\/2408.17231https:\/\/arxiv.org\/abs\/2408.17231"},{"key":"e_1_3_3_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300780"},{"key":"e_1_3_3_1_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"e_1_3_3_1_39_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-72390-2_49"},{"key":"e_1_3_3_1_40_1","doi-asserted-by":"publisher","unstructured":"Rakshit Kothari Zhizhuo Yang Christopher Kanan Reynold Bailey Jeff\u00a0B Pelz and Gabriel\u00a0J Diaz. 2020. Gaze-in-wild: A dataset for studying eye and head coordination in everyday activities. Scientific Reports 10 1 (2020) 1\u201318. 10.1038\/s41598-020-59251-5","DOI":"10.1038\/s41598-020-59251-5"},{"key":"e_1_3_3_1_41_1","doi-asserted-by":"publisher","unstructured":"Rakshit\u00a0S. Kothari A.\u00a0K. Chaudhary R.\u00a0J. Bailey J.\u00a0B. Pelz and G.\u00a0J. Diaz. 2021. EllSeg: An Ellipse Segmentation Framework for Robust Gaze Tracking. IEEE Transactions on Visualization & Computer Graphics 27 05 (5 2021) 2757\u20132767. 10.1109\/TVCG.2021.3067765","DOI":"10.1109\/TVCG.2021.3067765"},{"key":"e_1_3_3_1_42_1","doi-asserted-by":"publisher","unstructured":"Mikhail Kulyabin Aleksei Zhdanov Andrey Pershin Gleb Sokolov Anastasia Nikiforova Mikhail Ronkin Vasilii Borisov and Andreas Maier. 2024. Segment Anything in Optical Coherence Tomography: SAM 2 for Volumetric Segmentation of Retinal Biomarkers. Bioengineering 11 9 (2024). 10.3390\/bioengineering11090940","DOI":"10.3390\/bioengineering11090940"},{"key":"e_1_3_3_1_43_1","first-page":"79","volume-title":"Computer Vision and Pattern Recognition-Workshops, 2005. CVPR Workshops. IEEE Computer Society Conference on","author":"Li Dongheng","year":"2005","unstructured":"Dongheng Li, David Winfield, and Derrick\u00a0J Parkhurst. 2005. Starburst: A hybrid algorithm for video-based eye tracking combining feature-based and model-based approaches. In Computer Vision and Pattern Recognition-Workshops, 2005. CVPR Workshops. IEEE Computer Society Conference on. IEEE, 79\u201379."},{"key":"e_1_3_3_1_44_1","unstructured":"Haofeng Liu Erli Zhang Junde Wu Mingxuan Hong and Yueming Jin. 2024. Surgical SAM 2: Real-time Segment Anything in Surgical Video by Efficient Frame Pruning. arXiv abs\/2408.07931 (2024). arXiv:https:\/\/arXiv.org\/abs\/2408.07931https:\/\/arxiv.org\/abs\/2408.07931"},{"key":"e_1_3_3_1_45_1","doi-asserted-by":"crossref","unstructured":"Jun Ma Yuting He Feifei Li Lin Han Chenyu You and Bo Wang. 2024a. Segment anything in medical images. Nature Communications 15 1 (2024) 654.","DOI":"10.1038\/s41467-024-44824-z"},{"key":"e_1_3_3_1_46_1","unstructured":"Jun Ma Sumin Kim Feifei Li Mohammed Baharoon Reza Asakereh Hongwei Lyu and Bo Wang. 2024b. Segment Anything in Medical Images and Videos: Benchmark and Deployment. arXiv abs\/2408.03322 (2024). arXiv:https:\/\/arXiv.org\/abs\/2408.03322https:\/\/arxiv.org\/abs\/2408.03322"},{"key":"e_1_3_3_1_47_1","doi-asserted-by":"publisher","unstructured":"Virmarie Maquiling Sean\u00a0Anthony Byrne Diederick\u00a0C. Niehorster Marco Carminati and Enkelejda Kasneci. 2025. Zero-Shot Pupil Segmentation with SAM 2: A Case Study of Over 14 Million Images. Proc. ACM Comput. Graph. Interact. Tech. 8 2 Article 23 (5 2025) 16\u00a0pages. 10.1145\/3729409","DOI":"10.1145\/3729409"},{"key":"e_1_3_3_1_48_1","doi-asserted-by":"publisher","unstructured":"Virmarie Maquiling Sean\u00a0Anthony Byrne Diederick\u00a0C. Niehorster Marcus Nystr\u00f6m and Enkelejda Kasneci. 2024. Zero-Shot Segmentation of Eye Features Using the Segment Anything Model (SAM). Proc. ACM Comput. Graph. Interact. Tech. 7 2 Article 26 (5 2024) 16\u00a0pages. 10.1145\/3654704","DOI":"10.1145\/3654704"},{"key":"e_1_3_3_1_49_1","doi-asserted-by":"crossref","unstructured":"Maciej\u00a0A Mazurowski Haoyu Dong Hanxue Gu Jichen Yang Nicholas Konz and Yixin Zhang. 2023. Segment anything model for medical image analysis: an experimental study. Medical Image Analysis 89 (2023) 102918.","DOI":"10.1016\/j.media.2023.102918"},{"key":"e_1_3_3_1_50_1","volume-title":"The oculometer","author":"Merchant John","year":"1967","unstructured":"John Merchant. 1967. The oculometer. Technical Report CR-805. NASA, Washington, DC."},{"key":"e_1_3_3_1_51_1","doi-asserted-by":"crossref","unstructured":"John Merchant Richard Morrissette and James\u00a0L Porterfield. 1974. Remote measurement of eye direction allowing subject motion over one cubic foot of space. IEEE Transactions on Biomedical Engineering4 (1974) 309\u2013317.","DOI":"10.1109\/TBME.1974.324318"},{"key":"e_1_3_3_1_52_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2009.5414530"},{"key":"e_1_3_3_1_53_1","doi-asserted-by":"publisher","unstructured":"Diederick\u00a0C Niehorster Roy\u00a0S Hessels and Jeroen\u00a0S Benjamins. 2020a. GlassesViewer: Open-source software for viewing and analyzing data from the Tobii Pro Glasses 2 eye tracker. Behavior Research Methods 52 3 (2020) 1244\u20131253. 10.3758\/s13428-019-01314-1","DOI":"10.3758\/s13428-019-01314-1"},{"key":"e_1_3_3_1_54_1","doi-asserted-by":"publisher","unstructured":"Diederick\u00a0C Niehorster Thiago Santini Roy\u00a0S Hessels Ignace T\u00a0C Hooge Enkelejda Kasneci and Marcus Nystr\u00f6m. 2020b. The impact of slippage on the data quality of head-worn eye trackers. Behavior Research Methods 52 3 (2020) 1140\u20131160. 10.3758\/s13428-019-01307-0","DOI":"10.3758\/s13428-019-01307-0"},{"key":"e_1_3_3_1_55_1","doi-asserted-by":"publisher","unstructured":"Diederick\u00a0C Niehorster Raimondas Zemblys Tanya Beelders and Kenneth Holmqvist. 2020c. Characterizing gaze position signals and synthesizing noise during fixations in eye-tracking data. Behavior Research Methods 52 6 (2020) 2515\u20132534. 10.3758\/s13428-020-01400-9","DOI":"10.3758\/s13428-020-01400-9"},{"key":"e_1_3_3_1_56_1","doi-asserted-by":"publisher","unstructured":"Marcus Nystr\u00f6m Ignace T.\u00a0C. Hooge Roy\u00a0S. Hessels Richard Andersson Dan\u00a0Witzner Hansen Roger Johansson and Diederick\u00a0C. Niehorster. 2025. The fundamentals of eye tracking part 3: How to choose an eye tracker. Behavior Research Methods 57 2 (22 Jan 2025) 67. 10.3758\/s13428-024-02587-x","DOI":"10.3758\/s13428-024-02587-x"},{"key":"e_1_3_3_1_57_1","doi-asserted-by":"crossref","unstructured":"Marcus Nystr\u00f6m Diederick\u00a0C Niehorster Richard Andersson Roy\u00a0S Hessels and Ignace T\u00a0C Hooge. 2023. The amplitude of small eye movements can be accurately estimated with video-based eye trackers. Behavior Research Methods 55 2 (2023) 657\u2013669.","DOI":"10.3758\/s13428-021-01780-6"},{"key":"e_1_3_3_1_58_1","doi-asserted-by":"publisher","unstructured":"Marcus Nystr\u00f6m Ignace T\u00a0C Hooge and Kenneth Holmqvist. 2013. Post-saccadic oscillations in eye movement data recorded with pupil-based eye trackers reflect motion of the pupil inside the iris. Vision Research 92 (2013) 59\u201366. 10.1016\/j.visres.2013.09.009","DOI":"10.1016\/j.visres.2013.09.009"},{"key":"e_1_3_3_1_59_1","doi-asserted-by":"crossref","unstructured":"Cristina Palmero Abhishek Sharma Karsten Behrendt Kapil Krishnakumar Oleg\u00a0V Komogortsev and Sachin\u00a0S Talathi. 2021. OpenEDS2020 Challenge on Gaze Tracking for VR: Dataset and Results. Sensors 21 14 (2021) 4769.","DOI":"10.3390\/s21144769"},{"key":"e_1_3_3_1_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/3204493.3204545"},{"key":"e_1_3_3_1_61_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW53098.2021.00350"},{"key":"e_1_3_3_1_62_1","unstructured":"Alec Radford. 2018. Improving language understanding by generative pre-training. (2018)."},{"key":"e_1_3_3_1_63_1","unstructured":"Osher Rafaeli Tal Svoray Roni Blushtein-Livnon and Ariel Nahlieli. 2025. Prompt-Based Segmentation at Multiple Resolutions and Lighting Conditions using Segment Anything Model 2. arXiv abs\/2408.06970 (2025). arXiv:https:\/\/arXiv.org\/abs\/2408.06970https:\/\/arxiv.org\/abs\/2408.06970"},{"key":"e_1_3_3_1_64_1","unstructured":"Frano Raji\u010d Lei Ke Yu-Wing Tai Chi-Keung Tang Martin Danelljan and Fisher Yu. 2023. Segment Anything Meets Point Tracking. arXiv:https:\/\/arXiv.org\/abs\/2307.01197https:\/\/arxiv.org\/abs\/2307.01197"},{"key":"e_1_3_3_1_65_1","unstructured":"Nikhila Ravi Valentin Gabeur Yuan-Ting Hu Ronghang Hu Chaitanya Ryali Tengyu Ma Haitham Khedr Roman R\u00e4dle Chloe Rolland Laura Gustafson Eric Mintun Junting Pan Kalyan\u00a0Vasudev Alwala Nicolas Carion Chao-Yuan Wu Ross Girshick Piotr Doll\u00e1r and Christoph Feichtenhofer. 2024. SAM 2: Segment Anything in Images and Videos. arXiv abs\/2408.00714 (2024). arXiv:https:\/\/arXiv.org\/abs\/2408.00714http:\/\/arxiv.org\/abs\/2408.00714"},{"key":"e_1_3_3_1_66_1","doi-asserted-by":"publisher","unstructured":"Thiago Santini Wolfgang Fuhl and Enkelejda Kasneci. 2018a. PuRe: Robust pupil detection for real-time pervasive eye tracking. Computer Vision and Image Understanding 170 (2018) 40 \u2013 50. 10.1016\/j.cviu.2018.02.002","DOI":"10.1016\/j.cviu.2018.02.002"},{"key":"e_1_3_3_1_67_1","doi-asserted-by":"publisher","DOI":"10.1145\/3204493.3204578"},{"key":"e_1_3_3_1_68_1","unstructured":"Anna\u00a0Christina Schaar Alejandro Tejada-Lapuerta Giovanni Palla Robert Gutgesell Lennard Halle Mariia Minaeva Larsen Vornholz Leander Dony Francesca Drummer Mojtaba Bahrami et\u00a0al. 2024. Nicheformer: a foundation model for single-cell and spatial omics. bioRxiv (2024) 2024\u201304."},{"key":"e_1_3_3_1_69_1","doi-asserted-by":"crossref","unstructured":"Siddharth Shankar Leigh\u00a0A Stearns and CJ van\u00a0der Veen. 2023. Semantic segmentation of glaciological features across multiple remote sensing platforms with the Segment Anything Model (SAM). Journal of Glaciology (2023) 1\u201310.","DOI":"10.1017\/jog.2023.95"},{"key":"e_1_3_3_1_70_1","unstructured":"Qiuhong Shen Xingyi Yang and Xinchao Wang. 2023. Anything-3d: Towards single-view anything reconstruction in the wild. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2304.10261 (2023)."},{"key":"e_1_3_3_1_71_1","doi-asserted-by":"publisher","unstructured":"Dave\u00a0M Stampe. 1993. Heuristic filtering and reliable calibration methods for video-based pupil-tracking systems. Behavior Research Methods Instruments & Computers 25 2 (1993) 137\u2013142. 10.3758\/BF03204486","DOI":"10.3758\/BF03204486"},{"key":"e_1_3_3_1_72_1","doi-asserted-by":"publisher","DOI":"10.1145\/2168556.2168585"},{"key":"e_1_3_3_1_73_1","doi-asserted-by":"publisher","DOI":"10.1145\/2857491.2857520"},{"key":"e_1_3_3_1_74_1","unstructured":"Di Wang Jing Zhang Bo Du Minqiang Xu Lin Liu Dacheng Tao and Liangpei Zhang. 2024. Samrs: Scaling-up remote sensing segmentation dataset with segment anything model. Advances in Neural Information Processing Systems 36 (2024)."},{"key":"e_1_3_3_1_75_1","unstructured":"Cheng-Yen Yang Hsiang-Wei Huang Wenhao Chai Zhongyu Jiang and Jenq-Neng Hwang. 2024. SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking with Motion-Aware Memory. arXiv:https:\/\/arXiv.org\/abs\/2411.11922https:\/\/arxiv.org\/abs\/2411.11922"},{"key":"e_1_3_3_1_76_1","unstructured":"Jinyu Yang Mingqi Gao Zhe Li Shang Gao Fangjing Wang and Feng Zheng. 2023. Track Anything: Segment Anything Meets Videos. arXiv:https:\/\/arXiv.org\/abs\/2304.11968https:\/\/arxiv.org\/abs\/2304.11968"},{"key":"e_1_3_3_1_77_1","doi-asserted-by":"crossref","unstructured":"Yuk-Hoi Yiu Moustafa Aboulatta Theresa Raiser Leoni Ophey Virginia\u00a0L Flanagin Peter Zu\u00a0Eulenburg and Seyed-Ahmad Ahmadi. 2019. DeepVOG: Open-source pupil segmentation and gaze estimation in neuroscience using deep learning. Journal of Neuroscience Methods 324 (2019) 108307.","DOI":"10.1016\/j.jneumeth.2019.05.016"},{"key":"e_1_3_3_1_78_1","unstructured":"Chaoning Zhang Dongshen Han Yu Qiao Jung\u00a0Uk Kim Sung-Ho Bae Seungkyu Lee and Choong\u00a0Seon Hong. 2023. Faster Segment Anything: Towards Lightweight SAM for Mobile Applications. arXiv abs\/2306.14289 (2023). arXiv:https:\/\/arXiv.org\/abs\/2306.14289https:\/\/arxiv.org\/abs\/2306.14289"},{"key":"e_1_3_3_1_79_1","unstructured":"Tianfei Zhou Wang Xia Fei Zhang Boyu Chang Wenguan Wang Ye Yuan Ender Konukoglu and Daniel Cremers. 2024b. Image Segmentation in Foundation Model Era: A Survey. arXiv abs\/2408.12957 (2024). arXiv:https:\/\/arXiv.org\/abs\/2408.12957https:\/\/arxiv.org\/abs\/2408.12957"},{"key":"e_1_3_3_1_80_1","doi-asserted-by":"publisher","unstructured":"Yukun Zhou Mark\u00a0A. Chia Siegfried\u00a0K. Wagner Murat\u00a0S. Ayhan Dominic\u00a0J. Williamson Robbert\u00a0R. Struyven Timing Liu Moucheng Xu Mateo\u00a0G. Lozano Peter Woodward-Court Yuka Kihara UK Biobank Eye & Vision Consortium Naomi Allen John E.\u00a0J. Gallacher Thomas Littlejohns Tariq Aslam Paul Bishop Graeme Black Panagiotis Sergouniotis Denize Atan Andrew\u00a0D. Dick Cathy Williams Sarah Barman Jenny\u00a0H. Barrett Sarah Mackie Tasanee Braithwaite Roxana\u00a0O. Carare Sarah Ennis Jane Gibson Andrew\u00a0J. Lotery Jay Self Usha Chakravarthy Ruth\u00a0E. Hogg Euan Paterson Jayne Woodside Tunde Peto Gareth Mckay Bernadette Mcguinness Paul\u00a0J. Foster Konstantinos Balaskas Anthony\u00a0P. Khawaja Nikolas Pontikos Jugnoo\u00a0S. Rahi Gerassimos Lascaratos Praveen\u00a0J. Patel Michelle Chan Sharon Y.\u00a0L. Chua Alexander Day Parul Desai Cathy Egan Marcus Fruttiger David\u00a0F. Garway-Heath Alison Hardcastle Sir Peng\u00a0T. Khaw Tony Moore Sobha Sivaprasad Nicholas Strouthidis Dhanes Thomas Adnan Tufail Ananth\u00a0C. Viswanathan Bal Dhillon Tom Macgillivray Cathie Sudlow Veronique Vitart Alexander Doney Emanuele Trucco Jeremy\u00a0A. Guggeinheim James\u00a0E. Morgan Chris\u00a0J. Hammond Katie Williams Pirro Hysi Simon\u00a0P. Harding Yalin Zheng Robert Luben Phil Luthert Zihan Sun Martin McKibbin Eoin O\u2019Sullivan Richard Oram Mike Weedon Chris\u00a0G. Owen Alicja\u00a0R. Rudnicka Naveed Sattar David Steel Irene Stratton Robyn Tapp Max\u00a0M. Yates Axel Petzold Savita Madhusudhan Andre Altmann Aaron\u00a0Y. Lee Eric\u00a0J. Topol Alastair\u00a0K. Denniston Daniel\u00a0C. Alexander and Pearse\u00a0A. Keane. 2023. A foundation model for generalizable disease detection from retinal images. Nature 622 7981 (01 10 2023) 156\u2013163. 10.1038\/s41586-023-06555-x","DOI":"10.1038\/s41586-023-06555-x"},{"key":"e_1_3_3_1_81_1","unstructured":"Yuli Zhou Guolei Sun Yawei Li Luca Benini and Ender Konukoglu. 2024a. When SAM2 Meets Video Camouflaged Object Segmentation: A Comprehensive Evaluation and Adaptation. arXiv abs\/2409.18653 (2024). arXiv:https:\/\/arXiv.org\/abs\/2409.18653https:\/\/arxiv.org\/abs\/2409.18653"},{"key":"e_1_3_3_1_82_1","unstructured":"Jiayuan Zhu Abdullah Hamdi Yunli Qi Yueming Jin and Junde Wu. 2024. Medical SAM 2: Segment medical images as video via Segment Anything Model 2. arXiv abs\/2408.00874 (2024). arXiv:https:\/\/arXiv.org\/abs\/2408.00874https:\/\/arxiv.org\/abs\/2408.00874"}],"event":{"name":"ETRA '25: 2025 Symposium on Eye Tracking Research and Applications","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction","SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques"],"location":"Tokyo Japan","acronym":"ETRA '25"},"container-title":["Proceedings of the 2025 Symposium on Eye Tracking Research and Applications"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3715669.3723118","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:18:19Z","timestamp":1750295899000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3715669.3723118"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,25]]},"references-count":81,"alternative-id":["10.1145\/3715669.3723118","10.1145\/3715669"],"URL":"https:\/\/doi.org\/10.1145\/3715669.3723118","relation":{},"subject":[],"published":{"date-parts":[[2025,5,25]]},"assertion":[{"value":"2025-05-25","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}