{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T14:47:47Z","timestamp":1742914067048,"version":"3.40.3"},"publisher-location":"Cham","reference-count":32,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783031105210"},{"type":"electronic","value":"9783031105227"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-10522-7_26","type":"book-chapter","created":{"date-parts":[[2022,7,14]],"date-time":"2022-07-14T05:14:49Z","timestamp":1657775689000},"page":"373-384","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Eye Centre Localisation with Convolutional Neural Networks in High- and Low-Resolution Images"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8337-2595","authenticated-orcid":false,"given":"Wenhao","family":"Zhang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5307-8288","authenticated-orcid":false,"given":"Melvyn L.","family":"Smith","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,7,15]]},"reference":[{"issue":"1","key":"26_CR1","doi-asserted-by":"publisher","first-page":"21","DOI":"10.1068\/p080021","volume":"8","author":"K Rayner","year":"1979","unstructured":"Rayner, K.: Eye guidance in reading: fixation locations within words. Perception 8(1), 21\u201330 (1979)","journal-title":"Perception"},{"issue":"6","key":"26_CR2","doi-asserted-by":"publisher","first-page":"578","DOI":"10.3758\/BF03203972","volume":"17","author":"GW McConkie","year":"1975","unstructured":"McConkie, G.W., Rayner, K.: The span of the effective stimulus during a fixation in reading. Percept. Psychophys. 17(6), 578\u2013586 (1975)","journal-title":"Percept. Psychophys."},{"key":"26_CR3","unstructured":"Tobii pro Homepage. https:\/\/www.tobiipro.com\/. Accessed 02 Apr 2022"},{"key":"26_CR4","unstructured":"Gazepoint Homepage. https:\/\/www.gazept.com\/. Accessed 08 Mar 2022"},{"issue":"4","key":"26_CR5","doi-asserted-by":"publisher","first-page":"455","DOI":"10.3758\/BF03195475","volume":"34","author":"AT Duchowski","year":"2002","unstructured":"Duchowski, A.T.: A breadth-first survey of eye-tracking applications. Behav. Res. Methods Instrum. Comput. 34(4), 455\u2013470 (2002)","journal-title":"Behav. Res. Methods Instrum. Comput."},{"issue":"6","key":"26_CR6","first-page":"39","volume":"34","author":"DM Krugman","year":"1994","unstructured":"Krugman, D.M., Fox, R.J., Fletcher, J.E., Rojas, T.H.: Do adolescents attend to warnings in cigarette advertising? An eye-tracking approach. J. Advert. Res. 34(6), 39\u201353 (1994)","journal-title":"J. Advert. Res."},{"issue":"5","key":"26_CR7","doi-asserted-by":"publisher","first-page":"708","DOI":"10.1002\/acp.1742","volume":"25","author":"G Hervet","year":"2011","unstructured":"Hervet, G., Gu\u00e9rard, K., Tremblay, S., Chtourou, M.S.: Is banner blindness genuine? Eye tracking internet text advertising. Appl. Cogn. Psychol. 25(5), 708\u2013716 (2011)","journal-title":"Appl. Cogn. Psychol."},{"key":"26_CR8","doi-asserted-by":"publisher","first-page":"118","DOI":"10.3389\/fnagi.2015.00118","volume":"7","author":"TJ Crawford","year":"2015","unstructured":"Crawford, T.J., Devereaux, A., Higham, S., Kelly, C.: The disengagement of visual attention in Alzheimer\u2019s disease: a longitudinal eye-tracking study. Front. Aging Neurosci. 7, 118 (2015)","journal-title":"Front. Aging Neurosci."},{"key":"26_CR9","doi-asserted-by":"crossref","unstructured":"Kiili, K., Ketamo, H., Kickmeier-Rust, M.D.: Evaluating the usefulness of eye tracking in game-based learning. Int. J. Ser Games 1(2), 51\u201365 (2014)","DOI":"10.17083\/ijsg.v1i2.15"},{"key":"26_CR10","doi-asserted-by":"crossref","unstructured":"Zhang, X., Liu, X., Yuan, S.M., Lin, S.F.: Eye tracking based control system for natural human-computer interaction. Comput. Intell. Neurosci. 2017, 1\u20139 (2017)","DOI":"10.1155\/2017\/5739301"},{"issue":"1","key":"26_CR11","doi-asserted-by":"publisher","first-page":"261","DOI":"10.1007\/s10339-012-0499-z","volume":"13","author":"ML Mele","year":"2012","unstructured":"Mele, M.L., Federici, S.: Gaze and eye-tracking solutions for psychological research. Cogn. Process. 13(1), 261\u2013265 (2012)","journal-title":"Cogn. Process."},{"key":"26_CR12","unstructured":"Tobii pro nano. https:\/\/www.tobiipro.com\/product-listing\/nano\/. Accessed 08 Mar 2022"},{"issue":"1","key":"26_CR13","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/1743-0003-9-82","volume":"9","author":"M Gneo","year":"2012","unstructured":"Gneo, M., Schmid, M., Conforto, S., D\u2019Alessio, T.: A free geometry model-independent neural eye-gaze tracking system. J. Neuroeng. Rehabil. 9(1), 1\u201315 (2012)","journal-title":"J. Neuroeng. Rehabil."},{"key":"26_CR14","doi-asserted-by":"crossref","unstructured":"Binaee, K., Sinnott, C., Capurro, K.J., MacNeilage, P., Lescroart, M.D.: Pupil Tracking under direct sunlight. In: ACM Symposium on Eye Tracking Research and Applications, pp. 1\u20134. Association for Computing Machinery, New York (2021)","DOI":"10.1145\/3450341.3458490"},{"key":"26_CR15","first-page":"125","volume":"11","author":"F Timm","year":"2011","unstructured":"Timm, F., Barth, E.: Accurate eye centre localisation by means of gradients. Visapp 11, 125\u2013130 (2011)","journal-title":"Visapp"},{"key":"26_CR16","doi-asserted-by":"crossref","unstructured":"Villanueva, A., Ponz, V., Sesma-Sanchez, L., Ariz, M., Porta, S., Cabeza, R.: Hybrid method based on topography for robust detection of iris center and eye corners. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 9(4), 1\u201320 (2013)","DOI":"10.1145\/2501643.2501647"},{"issue":"7","key":"26_CR17","doi-asserted-by":"publisher","first-page":"660","DOI":"10.1049\/iet-cvi.2015.0316","volume":"10","author":"A George","year":"2016","unstructured":"George, A., Routray, A.: Fast and accurate algorithm for eye localisation for gaze tracking in low-resolution images. IET Comput. Vision 10(7), 660\u2013669 (2016)","journal-title":"IET Comput. Vision"},{"key":"26_CR18","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1016\/j.cviu.2016.03.014","volume":"149","author":"W Zhang","year":"2016","unstructured":"Zhang, W., Smith, M.L., Smith, L.N., Farooq, A.: Gender and gaze gesture recognition for human-computer interaction. Comput. Vis. Image Underst. 149, 32\u201350 (2016)","journal-title":"Comput. Vis. Image Underst."},{"key":"26_CR19","doi-asserted-by":"crossref","unstructured":"Ahmad, N., Yadav, K.S., Ahmed, M., Laskar, R.H., Hossain, A.: An integrated approach for eye centre localization using deep networks and rectangular-intensity-gradient technique. J. King Saud Univ.-Comput. Inf. Sci. (2022)","DOI":"10.1016\/j.jksuci.2022.02.015"},{"issue":"13","key":"26_CR20","doi-asserted-by":"publisher","first-page":"3785","DOI":"10.3390\/s20133785","volume":"20","author":"W Khan","year":"2020","unstructured":"Khan, W., Hussain, A., Kuru, K., Al-Askar, H.: Pupil localisation and eye centre estimation using machine learning and computer vision. Sensors 20(13), 3785 (2020)","journal-title":"Sensors"},{"key":"26_CR21","first-page":"1755","volume":"10","author":"DE King","year":"2009","unstructured":"King, D.E.: Dlib-ml: a machine learning toolkit. J. Mach. Learn. Res. 10, 1755\u20131758 (2009)","journal-title":"J. Mach. Learn. Res."},{"key":"26_CR22","doi-asserted-by":"crossref","unstructured":"Zhang, W., Smith, M.: Eye centre localisation with convolutional neural network based regression. In 2019 IEEE 4th International Conference on Image, Vision and Computing, pp. 88\u201394. IEEE (2019)","DOI":"10.1109\/ICIVC47709.2019.8980972"},{"key":"26_CR23","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2818\u20132826. IEEE (2016)","DOI":"10.1109\/CVPR.2016.308"},{"key":"26_CR24","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778. IEEE (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"26_CR25","unstructured":"The BioID. Face database (2014). https:\/\/www.bioid.com\/About\/BioID-Face-Database. Accessed 05 Feb 2019"},{"key":"26_CR26","doi-asserted-by":"crossref","unstructured":"Jesorsky, O., Kirchberg, K.J., Frischholz, R.W.: Robust face detection using the Hausdorff distance. In: Bigun, J., Smeraldi, F. (eds.) Audio- and Video-Based Biometric Person Authentication. AVBPA 2001. Lecture Notes in Computer Science, vol. 2091. Springer, Heidelberg (2001). https:\/\/doi.org\/10.1007\/3-540-45344-X_14","DOI":"10.1007\/3-540-45344-X_14"},{"key":"26_CR27","doi-asserted-by":"crossref","unstructured":"Crutcher, M.D., Calhoun-Haney, R., Manzanares, C.M., Lah, J.J., Levey, A.I., Zola, S.M.: Eye tracking during a visual paired comparison task as a predictor of early dementia. Am. J. Alzheimer\u2019s Dis. Other Dement.\u00ae 24(3), 258\u2013266 (2009)","DOI":"10.1177\/1533317509332093"},{"issue":"1","key":"26_CR28","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/s41598-018-37186-2","volume":"9","author":"A Oyama","year":"2019","unstructured":"Oyama, A., et al.: Novel method for rapid assessment of cognitive impairment using high-performance eye-tracking technology. Sci. Rep. 9(1), 1\u20139 (2019)","journal-title":"Sci. Rep."},{"key":"26_CR29","doi-asserted-by":"crossref","unstructured":"Zoph, B., Vasudevan, V., Shlens, J., Le, Q.V.: Learning transferable architectures for scalable image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8697\u20138710. IEEE (2018)","DOI":"10.1109\/CVPR.2018.00907"},{"key":"26_CR30","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.C.: Mobilenetv2: inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520. IEEE (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"26_CR31","unstructured":"Tan, M., Le, Q.: Efficientnetv2: smaller models and faster training. In: International Conference on Machine Learning, pp. 10096\u201310106. PMLR (2021)"},{"issue":"3","key":"26_CR32","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O., et al.: Imagenet large scale visual recognition challenge. Int. J. Comput. Vision 115(3), 211\u2013252 (2015)","journal-title":"Int. J. Comput. Vision"}],"container-title":["Lecture Notes in Computer Science","Computational Science and Its Applications \u2013 ICCSA 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-10522-7_26","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,14]],"date-time":"2022-07-14T05:22:11Z","timestamp":1657776131000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-10522-7_26"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031105210","9783031105227"],"references-count":32,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-10522-7_26","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"15 July 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICCSA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Computational Science and Its Applications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Malaga","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Spain","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 July 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7 July 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iccsa2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/iccsa.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CyberChair 4","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"279","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"57","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"24","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"20% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.6","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"8.7","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"285 Workshop submission accepted out of 815 submissions","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}