{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T03:36:05Z","timestamp":1743046565504,"version":"3.40.3"},"publisher-location":"Cham","reference-count":28,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031714634"},{"type":"electronic","value":"9783031714641"}],"license":[{"start":{"date-parts":[[2024,11,13]],"date-time":"2024-11-13T00:00:00Z","timestamp":1731456000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,13]],"date-time":"2024-11-13T00:00:00Z","timestamp":1731456000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-71464-1_18","type":"book-chapter","created":{"date-parts":[[2024,11,12]],"date-time":"2024-11-12T17:50:16Z","timestamp":1731433816000},"page":"209-220","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["KD-Eye: Lightweight Pupil Segmentation for\u00a0Eye Tracking on\u00a0VR Headsets via\u00a0Knowledge Distillation"],"prefix":"10.1007","author":[{"given":"Yanlin","family":"Li","sequence":"first","affiliation":[]},{"given":"Ning","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Guangrong","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Yiran","family":"Shen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,13]]},"reference":[{"key":"18_CR1","doi-asserted-by":"publisher","unstructured":"Clay, V., K\u00f6nig, P., Koenig, S.: Eye tracking in virtual reality. J. Eye Movement Res. 12, April 2019. https:\/\/doi.org\/10.16910\/jemr.12.1.3","DOI":"10.16910\/jemr.12.1.3"},{"key":"18_CR2","doi-asserted-by":"crossref","unstructured":"Lohr, D., Griffith, H., Komogortsev, O.: Eye know you: Metric learning for end-to-end biometric authentication using eye movements from a longitudinal dataset, April 2021","DOI":"10.1109\/TBIOM.2022.3167633"},{"key":"18_CR3","doi-asserted-by":"publisher","first-page":"160","DOI":"10.1016\/j.lwt.2015.11.066","volume":"68","author":"D Oliveira","year":"2016","unstructured":"Oliveira, D., et al.: Consumers\u2019 attention to functional food labels: Insights from eye-tracking and change detection in a case study with probiotic milk. LWT Food Sci. Technol. 68, 160\u2013167 (2016)","journal-title":"LWT Food Sci. Technol."},{"key":"18_CR4","doi-asserted-by":"crossref","unstructured":"Gwizdka, J., Hosseini, R., Cole, M., Wang, S.: Temporal dynamics of eye-tracking and eeg during reading and relevance decisions. Journal of the Association for Information Science and Technology 68(10), 2299\u20132312","DOI":"10.1002\/asi.23904"},{"key":"18_CR5","doi-asserted-by":"crossref","unstructured":"Zaretskaya, N., Bause, J., Polimeni, J.R., Grassi, P.R., Scheffler, K., Bartels, A.: Eye-selective fmri activity in human primary visual cortex: Comparison between 3 t and 9.4 t, and effects across cortical depth. NeuroImage 220, 117078 (2020). https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1053811920305644","DOI":"10.1016\/j.neuroimage.2020.117078"},{"issue":"1","key":"18_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/s41467-020-18360-5","volume":"11","author":"N Valliappan","year":"2020","unstructured":"Valliappan, N., Dai, N., Steinberg, E., He, J., Rogers, K., Ramachandran, V., Xu, P., Shojaeizadeh, M., Guo, L., Kohlhoff, K., et al.: Accelerating eye movement research via accurate and affordable smartphone eye tracking. Nat. Commun. 11(1), 1\u201312 (2020)","journal-title":"Nat. Commun."},{"key":"18_CR7","unstructured":"Cheng, Y., Wang, H., Bao, Y., Lu, F.: Appearance-based gaze estimation with deep learning: A review and benchmark. arXiv preprint arXiv:2104.12668 (2021)"},{"key":"18_CR8","doi-asserted-by":"crossref","unstructured":"Morimoto, C.H., Mimica, M.R.M.: Eye gaze tracking techniques for interactive applications. Comput. Vis. Image Underst. 98(1), 4\u201324 (2005)","DOI":"10.1016\/j.cviu.2004.07.010"},{"key":"18_CR9","doi-asserted-by":"crossref","unstructured":"Wang, X., Zhang, J., Zhang, H., Zhao, S., Liu, H.: Vision-based gaze estimation: a review. IEEE Trans. Cogn. Dev. Syst. (2021)","DOI":"10.1109\/TCDS.2021.3066465"},{"key":"18_CR10","doi-asserted-by":"crossref","unstructured":"Kim, J., et al.: Nvgaze: an anatomically-informed dataset for low-latency, near-eye gaze estimation. In: Proceedings of the 2019 CHI Conference. Association for Computing Machinery, New York (2019). https:\/\/doi.org\/10.1145\/3290605.3300780","DOI":"10.1145\/3290605.3300780"},{"key":"18_CR11","doi-asserted-by":"publisher","unstructured":"Cherif, Z., Nait-Ali, A., Motsch, J., Krebs, M.: An adaptive calibration of an infrared light device used for gaze tracking. In: Proceedings of the 19th IEEE Instrumentation and Measurement Technology Conference, vol.\u00a02, pp. 1029\u20131033 (2002). https:\/\/doi.org\/10.1109\/IMTC.2002.1007096","DOI":"10.1109\/IMTC.2002.1007096"},{"key":"18_CR12","doi-asserted-by":"publisher","unstructured":"Wang, K., Ji, Q.: Real time eye gaze tracking with 3d deformable eye-face model. In: Proceedings of IEEE International Conference on Computer Vision (ICCV), pp. 1003\u20131011 (2017). https:\/\/doi.org\/10.1109\/ICCV.2017.114","DOI":"10.1109\/ICCV.2017.114"},{"issue":"6","key":"18_CR13","doi-asserted-by":"publisher","first-page":"1124","DOI":"10.1109\/TBME.2005.863952","volume":"53","author":"E Guestrin","year":"2006","unstructured":"Guestrin, E., Eizenman, M.: General theory of remote gaze estimation using the pupil center and corneal reflections. IEEE Trans. Biomed. Eng. 53(6), 1124\u20131133 (2006). https:\/\/doi.org\/10.1109\/TBME.2005.863952","journal-title":"IEEE Trans. Biomed. Eng."},{"key":"18_CR14","doi-asserted-by":"crossref","unstructured":"Zhang, X., Sugano, Y., Fritz, M., Bulling, A.: Appearance-based gaze estimation in the wild. In: Proceedings of IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4511\u20134520 (2015)","DOI":"10.1109\/CVPR.2015.7299081"},{"key":"18_CR15","doi-asserted-by":"crossref","unstructured":"Kwok, T.C., Kiefer, P., Schinazi, V.R., Adams, B., Raubal, M.: Gaze-guided narratives: adapting audio guide content to gaze in virtual and real environments. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, pp. 1\u201312 (2019)","DOI":"10.1145\/3290605.3300721"},{"key":"18_CR16","doi-asserted-by":"crossref","unstructured":"Mariakakis, A., Goel, M., Aumi, M.T.I., Patel, S.N., Wobbrock, J.O.: SwitchBack: using focus and saccade tracking to guide users\u2019 attention for mobile task resumption. In: Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems, pp. 2953\u20132962 (2015)","DOI":"10.1145\/2702123.2702539"},{"key":"18_CR17","doi-asserted-by":"crossref","unstructured":"Lan, G., Heit, B., Scargill, T., Gorlatova, M.: GazeGraph: Graph-based few-shot cognitive context sensing from human visual behavior. In: Proceedings of the ACM Conference on Embedded Networked Sensor Systems (SenSys). pp. 422\u2013435 (2020)","DOI":"10.1145\/3384419.3430774"},{"issue":"4","key":"18_CR18","first-page":"1","volume":"2","author":"N Srivastava","year":"2018","unstructured":"Srivastava, N., Newn, J., Velloso, E.: Combining low and mid-level gaze features for desktop activity recognition. In: Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 2(4), 1\u201327 (2018)","journal-title":"In: Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies"},{"issue":"4","key":"18_CR19","first-page":"1","volume":"38","author":"J Kim","year":"2019","unstructured":"Kim, J., et al.: Foveated AR: dynamically-foveated augmented reality display. ACM Trans. Graph. 38(4), 1\u201315 (2019)","journal-title":"ACM Trans. Graph."},{"key":"18_CR20","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-319-24574-4_28","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28"},{"issue":"4","key":"18_CR21","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1109\/TNSRE.2002.806829","volume":"10","author":"R Barea","year":"2002","unstructured":"Barea, R., Boquete, L., Mazo, M., Lopez, E.: System for assisted mobility using eye movements based on electrooculography. IEEE Trans. Neural Syst. Rehabil. Eng. 10(4), 209\u2013218 (2002). https:\/\/doi.org\/10.1109\/TNSRE.2002.806829","journal-title":"IEEE Trans. Neural Syst. Rehabil. Eng."},{"issue":"4","key":"18_CR22","doi-asserted-by":"publisher","first-page":"137","DOI":"10.1109\/TBMEL.1963.4322822","volume":"10","author":"DA Robinson","year":"1963","unstructured":"Robinson, D.A.: A method of measuring eye movement using a scieral search coil in a magnetic field. IEEE Trans. Bio-med. Electron. 10(4), 137\u2013145 (1963). https:\/\/doi.org\/10.1109\/TBMEL.1963.4322822","journal-title":"IEEE Trans. Bio-med. Electron."},{"key":"18_CR23","unstructured":"Zhao, G., et al.: Ev-eye: rethinking high-frequency eye tracking through the lenses of event cameras. In: Oh, A., Naumann, T., Globerson, A., Saenko, K., Hardt, M., Levine, S. (eds.) Advances in Neural Information Processing Systems, vol.\u00a036, pp. 62169\u201362182. Curran Associates, Inc. (2023)"},{"key":"18_CR24","doi-asserted-by":"crossref","unstructured":"Fuhl, W., Kasneci, G., Kasneci, E.: Teyed: over 20 million real-world eye images with pupil, eyelid, and iris 2d and 3d segmentations, 2d and 3d landmarks, 3d eyeball, gaze vector, and eye movement types. In: 2021 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), pp. 367\u2013375. IEEE (2021)","DOI":"10.1109\/ISMAR52148.2021.00053"},{"key":"18_CR25","doi-asserted-by":"publisher","first-page":"365","DOI":"10.1007\/978-3-030-58558-7_22","volume-title":"Computer Vision - ECCV 2020","author":"X Zhang","year":"2020","unstructured":"Zhang, X., Park, S., Beeler, T., Bradley, D., Tang, S., Hilliges, O.: Eth-xgaze: a large scale dataset for gaze estimation under extreme head pose and gaze variation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M. (eds.) Computer Vision - ECCV 2020, pp. 365\u2013381. Springer, Cham (2020)"},{"key":"18_CR26","doi-asserted-by":"crossref","unstructured":"Wang, K., Su, H., Ji, Q.: Neuro-inspired eye tracking with eye movement dynamics. In: Proceedings of IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9823\u20139832 (2019)","DOI":"10.1109\/CVPR.2019.01006"},{"key":"18_CR27","doi-asserted-by":"crossref","unstructured":"Rezatofighi, H., Tsoi, N., Gwak, J., Sadeghian, A., Reid, I., Savarese, S.: Generalized intersection over union: a metric and a loss for bounding box regression. In: Proceedings of IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 658\u2013666 (2019)","DOI":"10.1109\/CVPR.2019.00075"},{"key":"18_CR28","doi-asserted-by":"crossref","unstructured":"Chaudhary, A.K., et al.: RITnet: real-time semantic segmentation of the eye for gaze tracking. In: Proceedings of IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW) (2019)","DOI":"10.1109\/ICCVW.2019.00568"}],"container-title":["Lecture Notes in Computer Science","Wireless Artificial Intelligent Computing Systems and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-71464-1_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,12]],"date-time":"2024-11-12T18:04:53Z","timestamp":1731434693000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-71464-1_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,13]]},"ISBN":["9783031714634","9783031714641"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-71464-1_18","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,13]]},"assertion":[{"value":"13 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The datasets used in this manuscript are all publicly available. The corresponding repositories are all properly cited in the manuscript.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Data Availability Statement"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Disclosure of Interests"}},{"value":"WASA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Wireless Artificial Intelligent Computing Systems and Applications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Qingdao","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21 June 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 June 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"wasa2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/wasa-conference.org\/WASA2024\/index.html","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}