{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,28]],"date-time":"2026-03-28T16:45:09Z","timestamp":1774716309767,"version":"3.50.1"},"reference-count":32,"publisher":"MDPI AG","issue":"17","license":[{"start":{"date-parts":[[2020,8,25]],"date-time":"2020-08-25T00:00:00Z","timestamp":1598313600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"<jats:p>This study develops an eye tracking method for autostereoscopic three-dimensional (3D) display systems for use in various environments. The eye tracking-based autostereoscopic 3D display provides low crosstalk and high-resolution 3D image experience seamlessly without 3D eyeglasses by overcoming the viewing position restriction. However, accurate and fast eye position detection and tracking are still challenging, owing to the various light conditions, camera control, thick eyeglasses, eyeglass sunlight reflection, and limited system resources. This study presents a robust, automated algorithm and relevant systems for accurate and fast detection and tracking of eye pupil centers in 3D with a single visual camera and near-infrared (NIR) light emitting diodes (LEDs). Our proposed eye tracker consists of eye\u2013nose detection, eye\u2013nose shape keypoint alignment, a tracker checker, and tracking with NIR LED on\/off control. Eye\u2013nose detection generates facial subregion boxes, including the eyes and nose, which utilize an Error-Based Learning (EBL) method for the selection of the best learnt database (DB). After detection, the eye\u2013nose shape alignment is processed by the Supervised Descent Method (SDM) with Scale-invariant Feature Transform (SIFT). The aligner is content-aware in the sense that corresponding designated aligners are applied based on image content classification, such as the various light conditions and wearing eyeglasses. The conducted experiments on real image DBs yield promising eye detection and tracking outcomes, even in the presence of challenging conditions.<\/jats:p>","DOI":"10.3390\/s20174787","type":"journal-article","created":{"date-parts":[[2020,8,25]],"date-time":"2020-08-25T09:24:56Z","timestamp":1598347496000},"page":"4787","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["Content-Aware Eye Tracking for Autostereoscopic 3D Display"],"prefix":"10.3390","volume":"20","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7151-703X","authenticated-orcid":false,"given":"Dongwoo","family":"Kang","sequence":"first","affiliation":[{"name":"Multimedia Processing Lab, Samsung Advanced Institute of Technology, Suwon 16678, Korea"}]},{"given":"Jingu","family":"Heo","sequence":"additional","affiliation":[{"name":"Multimedia Processing Lab, Samsung Advanced Institute of Technology, Suwon 16678, Korea"}]}],"member":"1968","published-online":{"date-parts":[[2020,8,25]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"876","DOI":"10.1109\/JPROC.2017.2686445","article-title":"Flat Panel Light-Field 3-D Display: Concept, Design, Rendering, and Calibration","volume":"105","author":"Nam","year":"2017","journal-title":"Proc. IEEE"},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Lee, S., Park, J., Heo, J., Kang, B., Kang, D., Hwang, H., Lee, J., Choi, Y., Choi, K., and Nam, D. (2018). Autostereoscopic 3D display using directional subpixel rendering. Opt. Express, 26.","DOI":"10.1364\/OE.26.020233"},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Park, J., Nam, D., Choi, S.Y., Lee, J.H., Park, D.S., and Kim, C.Y. (2013). Light field rendering of multi-view contents for high density light field 3D display. SID Symposium Digest of Technical Papers, Wiley.","DOI":"10.1002\/j.2168-0159.2013.tb06300.x"},{"key":"ref_4","unstructured":"Cho, Y.H., and Nam, D.K. (2019). Content Visualizing Device and Method. (20190139298), U.S. Patent."},{"key":"ref_5","unstructured":"Martinez, L.A.V., and Orozoco, L.F.E. (2016). Head-up Display System Using Auto-Stereoscopy 3D Transparent Electronic Display. (20160073098), U.S. Patent."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"340","DOI":"10.1109\/TITS.2007.895298","article-title":"Real-time detection of driver cognitive distraction using support vector machines","volume":"8","author":"Liang","year":"2007","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"54","DOI":"10.1049\/htl.2017.0020","article-title":"Real-time eye tracking for the assessment of driver fatigue","volume":"5","author":"Xu","year":"2018","journal-title":"Healthc. Technol. Lett."},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Soccini, A.M. (2017, January 18\u201322). Gaze estimation based on head movements in virtual reality applications using deep learning. Proceedings of the IEEE Virtual Reality (VR), Los Angeles, CA, USA.","DOI":"10.1109\/VR.2017.7892352"},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"2577","DOI":"10.1109\/TCE.2010.5681143","article-title":"Gaze tracking system at a distance for controlling IPTV","volume":"56","author":"Lee","year":"2010","journal-title":"IEEE Trans. Consum. Electron."},{"key":"ref_10","unstructured":"Grip, M. (2010). Selecting Autofocus Area in an Image. (7 860 382), U.S. Patent."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"293","DOI":"10.1109\/TCE.2019.2912802","article-title":"GazeVisual: A practical software tool and web application for performance evaluation of eye tracking systems","volume":"65","author":"Kar","year":"2019","journal-title":"IEEE Trans. Consum. Electron."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"347","DOI":"10.1109\/TCE.2012.6227433","article-title":"Real-time eye gaze tracking for gaming design and consumer electronics systems","volume":"58","author":"Corcoran","year":"2012","journal-title":"IEEE Trans. Consum. Electron."},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"179","DOI":"10.1109\/TCE.2019.2899869","article-title":"Convolutional neural network implementation for eye-gaze estimation on low-quality consumer imaging systems","volume":"65","author":"Lemley","year":"2019","journal-title":"IEEE Trans. Consum. Electron."},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"4","DOI":"10.1016\/j.cviu.2004.07.010","article-title":"Eye gaze tracking techniques for interactive applications","volume":"98","author":"Morimoto","year":"2005","journal-title":"Comput. Vis. Image Underst."},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"2246","DOI":"10.1109\/TBME.2007.895750","article-title":"Novel eye gaze tracking techniques under natural head movement","volume":"54","author":"Zhu","year":"2007","journal-title":"IEEE Trans. Biomed. Eng."},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Wu, B., Iandola, F., Jin, P.H., and Keutzer, K. (2017, January 21\u201326). SqueezeDet: Unified, small, low power fully convolutional neural networks for real-time object detection for autonomous driving. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPRW.2017.60"},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Xiong, X., and De la Torre, F. (2013, January 25\u201327). Supervised descent method and its applications to face alignment. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA.","DOI":"10.1109\/CVPR.2013.75"},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"137","DOI":"10.1023\/B:VISI.0000013087.49260.fb","article-title":"Robust real-time face detection","volume":"57","author":"Viola","year":"2004","journal-title":"Int. J. Comput. Vision"},{"key":"ref_19","unstructured":"Viola, P., and Jones, M. (2001, January 8\u201314). Rapid object detection using a boosted cascade of simple features. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Kauai, HI, USA."},{"key":"ref_20","unstructured":"Freund, Y., and Schapire, R.E. (1996, January 3\u20136). Experiments with a new boosting algorithm. Proceedings of the International Conference on Machine Learning, Bari, Italy."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"121","DOI":"10.1109\/TPAMI.2017.2781233","article-title":"Hyperface: A deep multi-task learning framework for face detection, landmark localization, pose estimation, and gender recognition","volume":"41","author":"Ranjan","year":"2019","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"1499","DOI":"10.1109\/LSP.2016.2603342","article-title":"Joint face detection and alignment using multitask cascaded convolutional networks","volume":"23","author":"Zhang","year":"2016","journal-title":"IEEE Signal Process. Lett."},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"91","DOI":"10.1023\/B:VISI.0000029664.99615.94","article-title":"Distinctive image features from scale-invariant keypoints","volume":"60","author":"Lowe","year":"2004","journal-title":"Int. J. Comput. Vis."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"681","DOI":"10.1109\/34.927467","article-title":"Active appearance models","volume":"23","author":"Cootes","year":"2001","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"38","DOI":"10.1006\/cviu.1995.1004","article-title":"Active shape models-their training and application","volume":"61","author":"Cootes","year":"1995","journal-title":"Comput. Vis. Image Underst."},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Kang, D., Heo, J., Kang, B., and Nam, D. (2019). Pupil detection and tracking for AR 3D under various circumstances. Electron. Imaging, 2019.","DOI":"10.2352\/ISSN.2470-1173.2019.15.AVM-055"},{"key":"ref_27","doi-asserted-by":"crossref","first-page":"807","DOI":"10.1016\/j.imavis.2009.08.002","article-title":"Multi-pie","volume":"28","author":"Gross","year":"2010","journal-title":"Image Vis. Comput."},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"273","DOI":"10.1007\/BF00994018","article-title":"Support-vector networks","volume":"20","author":"Cortes","year":"1995","journal-title":"Mach. Learn."},{"key":"ref_29","unstructured":"Ahlberg, J. (2001). CANDIDE-3-An Updated Parameterized Face. Dept. of Electrical Engineering, Link\u00f6ping University. Tech. Rep. LiTHISY-R-2326."},{"key":"ref_30","first-page":"903","article-title":"3D Head Pose and Facial Expression Tracking using a Single Camera","volume":"16","author":"Terissi","year":"2010","journal-title":"J. Univers. Comput. Sci."},{"key":"ref_31","unstructured":"Dodgson, N.A. (2004). Variation and extrema of human interpupillary distance. Stereoscopic Displays and Virtual Reality Systems XI, International Society for Optics and Photonics."},{"key":"ref_32","unstructured":"Marcelo, B., Bertozzi, A.L., and Sapiro, G. (2001, January 8\u201314). Navier-stokes, fluid dynamics, and image and video inpainting. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Kauai, HI, USA."}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/20\/17\/4787\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T10:06:09Z","timestamp":1760177169000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/20\/17\/4787"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,8,25]]},"references-count":32,"journal-issue":{"issue":"17","published-online":{"date-parts":[[2020,9]]}},"alternative-id":["s20174787"],"URL":"https:\/\/doi.org\/10.3390\/s20174787","relation":{},"ISSN":["1424-8220"],"issn-type":[{"value":"1424-8220","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,8,25]]}}}