{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:19:56Z","timestamp":1750220396196,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":19,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,5,25]],"date-time":"2021-05-25T00:00:00Z","timestamp":1621900800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"NSF EPSCOR","award":["OIA-1920896"],"award-info":[{"award-number":["OIA-1920896"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,5,25]]},"DOI":"10.1145\/3450341.3458491","type":"proceedings-article","created":{"date-parts":[[2021,5,25]],"date-time":"2021-05-25T13:48:55Z","timestamp":1621950535000},"page":"1-4","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Characterizing the Performance of Deep Neural Networks for Eye-Tracking"],"prefix":"10.1145","author":[{"given":"Arnab","family":"Biswas","sequence":"first","affiliation":[{"name":"Psychology University of Nevada, Reno, United States"}]},{"given":"Kamran","family":"Binaee","sequence":"additional","affiliation":[{"name":"Psychology University of Nevada, Reno, United States"}]},{"given":"Kaylie Jacleen","family":"Capurro","sequence":"additional","affiliation":[{"name":"Psychology University of Nevada, United States"}]},{"given":"Mark D.","family":"Lescroart","sequence":"additional","affiliation":[{"name":"Psychology University of Nevada, Reno, United States"}]}],"member":"320","published-online":{"date-parts":[[2021,5,25]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Leonard Carmichael and Walter\u00a0F Dearborn. 1947. Reading and visual fatigue.Houghton Mifflin.  Leonard Carmichael and Walter\u00a0F Dearborn. 1947. Reading and visual fatigue.Houghton Mifflin."},{"key":"e_1_3_2_1_2_1","volume-title":"RITnet: Real-time Semantic Segmentation of the Eye for Gaze Tracking. In 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW). 3698\u20133702","author":"Chaudhary K.","year":"2019","unstructured":"A.\u00a0 K. Chaudhary , R. Kothari , M. Acharya , S. Dangi , N. Nair , R. Bailey , C. Kanan , G. Diaz , and J.\u00a0 B. Pelz . 2019 . RITnet: Real-time Semantic Segmentation of the Eye for Gaze Tracking. In 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW). 3698\u20133702 . https:\/\/doi.org\/10.1109\/ICCVW.2019.00568 A.\u00a0K. Chaudhary, R. Kothari, M. Acharya, S. Dangi, N. Nair, R. Bailey, C. Kanan, G. Diaz, and J.\u00a0B. Pelz. 2019. RITnet: Real-time Semantic Segmentation of the Eye for Gaze Tracking. In 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW). 3698\u20133702. https:\/\/doi.org\/10.1109\/ICCVW.2019.00568"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1364\/JOSA.48.000808"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/3314111.3319914"},{"key":"e_1_3_2_1_6_1","volume-title":"Deep Residual Learning for Image Recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 770\u2013778","author":"He K.","year":"2016","unstructured":"K. He , X. Zhang , S. Ren , and J. Sun . 2016 . Deep Residual Learning for Image Recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 770\u2013778 . https:\/\/doi.org\/10.1109\/CVPR. 2016 .90 K. He, X. Zhang, S. Ren, and J. Sun. 2016. Deep Residual Learning for Image Recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 770\u2013778. https:\/\/doi.org\/10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_1_7_1","volume-title":"imgaug. https:\/\/github.com\/aleju\/imgaug. Online","author":"Jung B.","year":"2020","unstructured":"Alexander\u00a0 B. Jung , Kentaro Wada , Jon Crall , Satoshi Tanaka , Jake Graving , Christoph Reinders , Sarthak Yadav , Joy Banerjee , G\u00e1bor Vecsei , Adam Kraft , Zheng Rui , Jirka Borovec , Christian Vallentin , Semen Zhydenko , Kilian Pfeiffer , Ben Cook , Ismael Fern\u00e1ndez , Fran\u00e7ois-Michel De\u00a0Rainville , Chi-Hung Weng , Abner Ayala-Acevedo , Raphael Meudec , Matias Laporte , 2020. imgaug. https:\/\/github.com\/aleju\/imgaug. Online ; accessed 01- Feb- 2020 . Alexander\u00a0B. Jung, Kentaro Wada, Jon Crall, Satoshi Tanaka, Jake Graving, Christoph Reinders, Sarthak Yadav, Joy Banerjee, G\u00e1bor Vecsei, Adam Kraft, Zheng Rui, Jirka Borovec, Christian Vallentin, Semen Zhydenko, Kilian Pfeiffer, Ben Cook, Ismael Fern\u00e1ndez, Fran\u00e7ois-Michel De\u00a0Rainville, Chi-Hung Weng, Abner Ayala-Acevedo, Raphael Meudec, Matias Laporte, 2020. imgaug. https:\/\/github.com\/aleju\/imgaug. Online; accessed 01-Feb-2020."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/2638728.2641695"},{"key":"e_1_3_2_1_9_1","unstructured":"Ryan Kiros Ruslan Salakhutdinov and Richard\u00a0S. Zemel. 2014. Unifying Visual-Semantic Embeddings with Multimodal Neural Language Models. CoRR abs\/1411.2539(2014). arxiv:1411.2539http:\/\/arxiv.org\/abs\/1411.2539  Ryan Kiros Ruslan Salakhutdinov and Richard\u00a0S. Zemel. 2014. Unifying Visual-Semantic Embeddings with Multimodal Neural Language Models. CoRR abs\/1411.2539(2014). arxiv:1411.2539http:\/\/arxiv.org\/abs\/1411.2539"},{"key":"e_1_3_2_1_10_1","volume-title":"Gaze-in-wild: A dataset for studying eye and head coordination in everyday activities. Scientific reports 10, 1","author":"Kothari Rakshit","year":"2020","unstructured":"Rakshit Kothari , Zhizhuo Yang , Christopher Kanan , Reynold Bailey , Jeff\u00a0 B Pelz , and Gabriel\u00a0 J Diaz . 2020 b. Gaze-in-wild: A dataset for studying eye and head coordination in everyday activities. Scientific reports 10, 1 (2020), 1\u201318. Rakshit Kothari, Zhizhuo Yang, Christopher Kanan, Reynold Bailey, Jeff\u00a0B Pelz, and Gabriel\u00a0J Diaz. 2020b. Gaze-in-wild: A dataset for studying eye and head coordination in everyday activities. Scientific reports 10, 1 (2020), 1\u201318."},{"key":"e_1_3_2_1_11_1","unstructured":"Rakshit\u00a0S Kothari Aayush\u00a0K Chaudhary Reynold\u00a0J Bailey Jeff\u00a0B Pelz and Gabriel\u00a0J Diaz. 2020a. EllSeg: An Ellipse Segmentation Framework for Robust Gaze Tracking. arXiv preprint arXiv:2007.09600(2020).  Rakshit\u00a0S Kothari Aayush\u00a0K Chaudhary Reynold\u00a0J Bailey Jeff\u00a0B Pelz and Gabriel\u00a0J Diaz. 2020a. EllSeg: An Ellipse Segmentation Framework for Robust Gaze Tracking. arXiv preprint arXiv:2007.09600(2020)."},{"key":"e_1_3_2_1_12_1","volume-title":"Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems 25","author":"Krizhevsky Alex","year":"2012","unstructured":"Alex Krizhevsky , Ilya Sutskever , and Geoffrey\u00a0 E Hinton . 2012. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems 25 ( 2012 ), 1097\u20131105. Alex Krizhevsky, Ilya Sutskever, and Geoffrey\u00a0E Hinton. 2012. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems 25 (2012), 1097\u20131105."},{"key":"e_1_3_2_1_13_1","volume-title":"DeepLabCut: markerless pose estimation of user-defined body parts with deep learning. Nature neuroscience 21, 9","author":"Mathis Alexander","year":"2018","unstructured":"Alexander Mathis , Pranav Mamidanna , Kevin\u00a0 M Cury , Taiga Abe , Venkatesh\u00a0 N Murthy , Mackenzie\u00a0Weygandt Mathis , and Matthias Bethge . 2018. DeepLabCut: markerless pose estimation of user-defined body parts with deep learning. Nature neuroscience 21, 9 ( 2018 ), 1281\u20131289. Alexander Mathis, Pranav Mamidanna, Kevin\u00a0M Cury, Taiga Abe, Venkatesh\u00a0N Murthy, Mackenzie\u00a0Weygandt Mathis, and Matthias Bethge. 2018. DeepLabCut: markerless pose estimation of user-defined body parts with deep learning. Nature neuroscience 21, 9 (2018), 1281\u20131289."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41596-019-0176-0"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.cell.2019.04.005"},{"key":"e_1_3_2_1_16_1","volume-title":"5555. High Speed and High Dynamic Range Video with an Event Camera","author":"Rebecq H.","year":"2019","unstructured":"H. Rebecq , R. Ranftl , V. Koltun , and D. Scaramuzza . 5555. High Speed and High Dynamic Range Video with an Event Camera . IEEE Transactions on Pattern Analysis & Machine Intelligence 01 (dec 5555), 1\u20131. https:\/\/doi.org\/10.1109\/TPAMI. 2019 .2963386 H. Rebecq, R. Ranftl, V. Koltun, and D. Scaramuzza. 5555. High Speed and High Dynamic Range Video with an Event Camera. IEEE Transactions on Pattern Analysis & Machine Intelligence01 (dec 5555), 1\u20131. https:\/\/doi.org\/10.1109\/TPAMI.2019.2963386"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2011.5995720"},{"key":"e_1_3_2_1_18_1","unstructured":"Grant Van\u00a0Horn and Pietro Perona. 2017. The devil is in the tails: Fine-grained classification in the wild. arXiv preprint arXiv:1709.01450(2017).  Grant Van\u00a0Horn and Pietro Perona. 2017. The devil is in the tails: Fine-grained classification in the wild. arXiv preprint arXiv:1709.01450(2017)."},{"key":"e_1_3_2_1_19_1","volume-title":"Translating Videos to Natural Language Using Deep Recurrent Neural Networks. (May\u2013June","author":"Venugopalan Subhashini","year":"2015","unstructured":"Subhashini Venugopalan , Huijuan Xu , Jeff Donahue , Marcus Rohrbach , Raymond Mooney , and Kate Saenko . 2015. Translating Videos to Natural Language Using Deep Recurrent Neural Networks. (May\u2013June 2015 ), 1494\u20131504. https:\/\/doi.org\/10.3115\/v1\/N15-1173 Subhashini Venugopalan, Huijuan Xu, Jeff Donahue, Marcus Rohrbach, Raymond Mooney, and Kate Saenko. 2015. Translating Videos to Natural Language Using Deep Recurrent Neural Networks. (May\u2013June 2015), 1494\u20131504. https:\/\/doi.org\/10.3115\/v1\/N15-1173"}],"event":{"name":"ETRA '21: 2021 Symposium on Eye Tracking Research and Applications","sponsor":["SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques"],"location":"Virtual Event Germany","acronym":"ETRA '21"},"container-title":["ACM Symposium on Eye Tracking Research and Applications"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3450341.3458491","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3450341.3458491","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:18:44Z","timestamp":1750191524000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3450341.3458491"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,5,25]]},"references-count":19,"alternative-id":["10.1145\/3450341.3458491","10.1145\/3450341"],"URL":"https:\/\/doi.org\/10.1145\/3450341.3458491","relation":{},"subject":[],"published":{"date-parts":[[2021,5,25]]},"assertion":[{"value":"2021-05-25","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}