{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T01:33:28Z","timestamp":1760232808019,"version":"build-2065373602"},"reference-count":24,"publisher":"MDPI AG","issue":"23","license":[{"start":{"date-parts":[[2022,12,2]],"date-time":"2022-12-02T00:00:00Z","timestamp":1669939200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"<jats:p>Eye-gaze direction-tracking technology is used in fields such as medicine, education, engineering, and gaming. Stability, accuracy, and precision of eye-gaze direction-tracking are demanded with simultaneous upgrades in response speed. In this study, a method is proposed to improve the speed with decreases in the system load and precision in the human pupil orbit model (HPOM) estimation method. The new method was proposed based on the phenomenon that the minor axis of the elliptical-deformed pupil always pointed toward the rotational center presented in various eye-gaze direction detection studies and HPOM estimation methods. Simulation experimental results confirmed that the speed was improved by at least 74 times by consuming less than 7 ms compared to the HPOM estimation. The accuracy of the eye\u2019s ocular rotational center point showed a maximum error of approximately 0.2 pixels on the x-axis and approximately 8 pixels on the y-axis. The precision of the proposed method was 0.0 pixels when the number of estimation samples (ES) was 7 or less, which showed results consistent with those of the HPOM estimation studies. However, the proposed method was judged to work conservatively against the allowable angle error (AAE), considering that the experiment was conducted under the worst conditions and the cost used to estimate the final model. Therefore, the proposed method could estimate HPOM with high accuracy and precision through AAE adjustment according to system performance and the usage environment.<\/jats:p>","DOI":"10.3390\/s22239398","type":"journal-article","created":{"date-parts":[[2022,12,2]],"date-time":"2022-12-02T03:28:04Z","timestamp":1669951684000},"page":"9398","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Improving Performance of the Human Pupil Orbit Model (HPOM) Estimation Method for Eye-Gaze Tracking"],"prefix":"10.3390","volume":"22","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3617-0525","authenticated-orcid":false,"given":"Seungbong","family":"Lee","sequence":"first","affiliation":[{"name":"Department of Medical Biotechnology, Dongguk University Biomedi Campus, 32, Dongguk-ro, Ilsan dong-gu, Goyang-si 10326, Republic of Korea"}]},{"given":"Jaehoon","family":"Jeong","sequence":"additional","affiliation":[{"name":"Medical Device Industry Program in Graduate School, Dongguk University, 30, Pildong-ro 1-gill, Jung-gu, Seoul 04620, Republic of Korea"}]},{"given":"Nahyun","family":"Kim","sequence":"additional","affiliation":[{"name":"Department of Medical Biotechnology, Dongguk University Biomedi Campus, 32, Dongguk-ro, Ilsan dong-gu, Goyang-si 10326, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2156-3905","authenticated-orcid":false,"given":"Manjae","family":"Shin","sequence":"additional","affiliation":[{"name":"Department of Medical Biotechnology, Dongguk University Biomedi Campus, 32, Dongguk-ro, Ilsan dong-gu, Goyang-si 10326, Republic of Korea"}]},{"given":"Sungmin","family":"Kim","sequence":"additional","affiliation":[{"name":"Department of Medical Biotechnology, Dongguk University Biomedi Campus, 32, Dongguk-ro, Ilsan dong-gu, Goyang-si 10326, Republic of Korea"}]}],"member":"1968","published-online":{"date-parts":[[2022,12,2]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Punde, P.A., Jadhav, M.E., and Manza, R.R. (2017, January 5\u20136). A study of eye tracking technology and its applications. Proceedings of the 2017 1st International Conference on Intelligent Systems and Information Management (ICISIM), Maharashtra, India.","DOI":"10.1109\/ICISIM.2017.8122153"},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Heldal, I., Helgesen, C., Ali, Q., Patel, D., Geitung, A.B., and Pettersen, H. (2021). Supporting School Aged Children to Train Their Vision by Using Serious Games. Computers, 10.","DOI":"10.3390\/computers10040053"},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"22402","DOI":"10.1109\/JSEN.2021.3105706","article-title":"Multi-sensor eye-tracking systems and tools for capturing Student attention and understanding engagement in learning: A review","volume":"21","author":"Wang","year":"2021","journal-title":"IEEE Sens. J."},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Gotoa, R., Kiyotab, K., Shimakawab, M., Watanabeb, K., and Okumab, C. (2021, January 26\u201331). Development of Training Game Application using Eye-gaze Control Technology to Support Employment of Physically challenged people. Proceedings of the 9th IIAE International Conference on Industrial Application Engineering 2021, Kitakyushu, Japan.","DOI":"10.12792\/iciae2021.026"},{"key":"ref_5","first-page":"4171","article-title":"Directions for 3D User Interface Research from Consumer VR Games","volume":"27","author":"Steed","year":"2021","journal-title":"Inst. Electr. Electron. Eng. Trans. Vis. Comput. Graph."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Smith, P., Dombrowski, M., McLinden, S., MacDonald, C., Lynn, D., Sparkman, J., Courbin, D., and Manero, A. (2022, January 8\u201311). Advancing dignity for adaptive wheelchair users via a hybrid eye tracking and electromyography training game. Proceedings of the 2022 Symposium on Eye Tracking Research and Applications June (ETRA), Seattle, WA, USA.","DOI":"10.1145\/3517031.3529612"},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Wang, Y., Ding, X., Yuan, G., and Fu, X. (2022). Dual-Cameras-Based Driver\u2019s Eye Gaze Tracking System with Non-Linear Gaze Point Refinement. Sensors, 22.","DOI":"10.3390\/s22062326"},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3491207","article-title":"The Eye in Extended Reality: A Survey on Gaze Interaction and Eye Tracking in Head-worn Extended Reality","volume":"55","author":"Plopski","year":"2022","journal-title":"ACM Comput. Surv."},{"key":"ref_9","unstructured":"Skaramagkas, V., Giannakakis, G., Ktistakis, E., Manousos, D., Karatzanis, I., Tachos, N., Tripoliti, E.E., Marias, K., Fotiadis, D.I., and Tsiknakis, M. (2021). Review of eye tracking metrics involved in emotional and cognitive processes. IEEE Rev. Biomed. Eng., 1."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"117","DOI":"10.1016\/bs.pbr.2019.04.036","article-title":"Translation and eccentric rotation in ocular motor modeling","volume":"248","author":"Demer","year":"2019","journal-title":"Prog. Brain Res."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"556441","DOI":"10.3389\/fneur.2020.556441","article-title":"Positional change of the eyeball during eye movements: Evidence of translatory movement","volume":"11","author":"Moon","year":"2020","journal-title":"Front. Neurol."},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Lee, S., Jeong, J., Kim, D., and Kim, S. (2022). Presenting a Human Pupil Orbit Model (HPOM) for Eye-Gaze Tracking. Appl. Sci., 12.","DOI":"10.3390\/app12168035"},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"23242","DOI":"10.1109\/ACCESS.2018.2828400","article-title":"A Geometry-Appearance-Based Pupil Detection Method for Near-Infrared Head-Mounted Cameras","volume":"6","author":"Li","year":"2018","journal-title":"IEEE Access"},{"key":"ref_14","unstructured":"Takegami, T., Gotoh, T., Kagei, S., and Tachino, R. (2003, January 10\u201312). A Hough Based Eye Direction Detection Algorithm without On-site Calibration. Proceedings of the 7th Digital Image Computing: Techniques and Applications (DICTA), Sydney, Australia."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Jin, N., Mavromatis, S., Sequeira, J., and Curcio, S. (2020). A Robust Method of Eye Torsion Measurement for Medical Applications. Information, 11.","DOI":"10.3390\/info11090408"},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"108307","DOI":"10.1016\/j.jneumeth.2019.05.016","article-title":"DeepVOG: Open-source pupil segmentation and gaze estimation in neuroscience using deep learning","volume":"324","author":"Yiu","year":"2019","journal-title":"J. Neurosci. Methods"},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"047003","DOI":"10.1117\/1.3562327","article-title":"Eye gaze estimation from the elliptical features of one iris","volume":"50","author":"Wen","year":"2011","journal-title":"Opt. Eng."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"9360","DOI":"10.1038\/s41598-019-45827-3","article-title":"A model of the entrance pupil of the human eye","volume":"9","author":"Aguirre","year":"2019","journal-title":"Sci. Rep."},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Imai, T., Takimoto, Y., Takeda, N., Uno, A., Inohara, H., and Shimada, S. (2016). High-Speed Video-Oculography for Measuring Three-Dimensional Rotation Vectors of Eye Movements in Mice. PLoS ONE, 11.","DOI":"10.1371\/journal.pone.0152307"},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"2575","DOI":"10.1007\/s00221-017-4990-8","article-title":"Three-dimensional analysis of linear vestibulo-ocular reflex in humans during eccentric rotation while facing downwards","volume":"235","author":"Imai","year":"2017","journal-title":"Exp. Brain Res."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"3","DOI":"10.1167\/13.6.3","article-title":"Pupil shape as viewed along the horizontal visual field","volume":"13","author":"Mathur","year":"2013","journal-title":"J. Vis."},{"key":"ref_22","unstructured":"Cai, W., Yu, Q., and Wang, H. (2004, January 15\u201319). A fast contour-based approach to circle and ellipse detection. Proceedings of the Fifth World Congress on Intelligent Control and Automation (IEEE Cat. No. 04EX788), Hangzhou, China."},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"205","DOI":"10.1137\/0702016","article-title":"Calculating the singular values and pseudo-inverse of a matrix","volume":"2","author":"Golub","year":"1965","journal-title":"J. Soc. Ind. Appl. Math. Ser. B Numer. Anal."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"381","DOI":"10.1145\/358669.358692","article-title":"Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography","volume":"24","author":"Fischler","year":"1981","journal-title":"Commun. ACM"}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/22\/23\/9398\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T01:32:32Z","timestamp":1760146352000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/22\/23\/9398"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,12,2]]},"references-count":24,"journal-issue":{"issue":"23","published-online":{"date-parts":[[2022,12]]}},"alternative-id":["s22239398"],"URL":"https:\/\/doi.org\/10.3390\/s22239398","relation":{},"ISSN":["1424-8220"],"issn-type":[{"type":"electronic","value":"1424-8220"}],"subject":[],"published":{"date-parts":[[2022,12,2]]}}}