{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,18]],"date-time":"2026-01-18T06:32:33Z","timestamp":1768717953734,"version":"3.49.0"},"reference-count":49,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2024,5,20]],"date-time":"2024-05-20T00:00:00Z","timestamp":1716163200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,5,20]],"date-time":"2024-05-20T00:00:00Z","timestamp":1716163200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001409","name":"Department of Science and Technology, Ministry of Science and Technology, India","doi-asserted-by":"publisher","award":["DST\/SEED\/TIDE\/2018\/66"],"award-info":[{"award-number":["DST\/SEED\/TIDE\/2018\/66"]}],"id":[{"id":"10.13039\/501100001409","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Real-Time Image Proc"],"published-print":{"date-parts":[[2024,6]]},"DOI":"10.1007\/s11554-024-01478-w","type":"journal-article","created":{"date-parts":[[2024,5,20]],"date-time":"2024-05-20T11:01:58Z","timestamp":1716202918000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["Empowering individuals with disabilities: a real-time, cost-effective, calibration-free assistive system utilizing eye tracking"],"prefix":"10.1007","volume":"21","author":[{"given":"Govind Ram","family":"Chhimpa","sequence":"first","affiliation":[]},{"given":"Ajay","family":"Kumar","sequence":"additional","affiliation":[]},{"given":"Sunita","family":"Garhwal","sequence":"additional","affiliation":[]},{"family":"Dhiraj","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,5,20]]},"reference":[{"issue":"4","key":"1478_CR1","doi-asserted-by":"publisher","first-page":"1003","DOI":"10.1007\/s10209-016-0493-9","volume":"16","author":"P Blignaut","year":"2017","unstructured":"Blignaut, P.: Development of a gaze-controlled support system for a person in an advanced stage of multiple sclerosis: a case study. Univ. Access Inf. Soc. 16(4), 1003\u20131016 (2017). https:\/\/doi.org\/10.1007\/s10209-016-0493-9","journal-title":"Univ. Access Inf. Soc."},{"key":"1478_CR2","doi-asserted-by":"publisher","first-page":"171","DOI":"10.4103\/2152-7806.169561","volume":"6","author":"S Zarei","year":"2015","unstructured":"Zarei, S., Carr, K., Reiley, L., Diaz, K., Guerra, O., Altamirano, P.F., Pagani, W., Lodin, D., Orozco, G., Chinea, A.: A comprehensive review of amyotrophic lateral sclerosis. Surg. Neurol. Int. 6, 171\u2013194 (2015). https:\/\/doi.org\/10.4103\/2152-7806.169561","journal-title":"Surg. Neurol. Int."},{"key":"1478_CR3","doi-asserted-by":"crossref","unstructured":"Mustaquim, M.: Gaze interaction\u2013a challenge for inclusive design. In: International Conference on Innovative Computing Technology, pp. 244\u2013250 (2011), Springer, Berlin, Heidelberg.","DOI":"10.1007\/978-3-642-27337-7_22"},{"issue":"8","key":"1478_CR4","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0159809","volume":"11","author":"N Saikia","year":"2016","unstructured":"Saikia, N., Bora, J.K., Jasilionis, D., Shkolnikov, V.M.: Disability divides in India: evidence from the 2011 census. PLoS ONE 11(8), e0159809 (2016)","journal-title":"PLoS ONE"},{"issue":"4","key":"1478_CR5","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1007\/s11554-023-01336-1","volume":"20","author":"GR Chhimpa","year":"2023","unstructured":"Chhimpa, G.R., Kumar, A., Garhwal, S., Dhiraj: Development of a real-time eye movement-based computer interface for communication with improved accuracy for disabled people under natural head movements. J. Real-Time Image Process. 20(4), 81 (2023). https:\/\/doi.org\/10.1007\/s11554-023-01336-1","journal-title":"J. Real-Time Image Process."},{"issue":"4","key":"1478_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/LSENS.2017.2720718","volume":"1","author":"T Wu","year":"2017","unstructured":"Wu, T., Wang, P., Lin, Y., Zhou, C.: A robust noninvasive eye control approach for disabled people based on Kinect 2.0 sensor. IEEE Sens. Lett. 1(4), 1\u20134 (2017). https:\/\/doi.org\/10.1109\/LSENS.2017.2720718","journal-title":"IEEE Sens. Lett."},{"issue":"2","key":"1478_CR7","doi-asserted-by":"publisher","first-page":"312","DOI":"10.1109\/THMS.2021.3123202","volume":"52","author":"M Porta","year":"2021","unstructured":"Porta, M., Dondi, P., Pianetta, A., Cantoni, V.: SPEye: a calibration-free gaze-driven text entry technique based on smooth pursuit. IEEE Trans. Human-Mach. Syst. 52(2), 312\u2013323 (2021). https:\/\/doi.org\/10.1109\/THMS.2021.3123202","journal-title":"IEEE Trans. Human-Mach. Syst."},{"key":"1478_CR8","doi-asserted-by":"crossref","unstructured":"Chew, M.T., Penver, K.: Low-cost eye gesture communication system for people with motor disabilities. In: 2019 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), pp. 1\u20135 (2019)","DOI":"10.1109\/I2MTC.2019.8826976"},{"issue":"4","key":"1478_CR9","doi-asserted-by":"publisher","first-page":"3125","DOI":"10.1007\/s10055-023-00858-0","volume":"27","author":"G S\u00e1nchez-Brizuela","year":"2023","unstructured":"S\u00e1nchez-Brizuela, G., Cisnal, A., de la Fuente-L\u00f3pez, E., Fraile, J.C., P\u00e9rez-Turiel, J.: Lightweight real-time hand segmentation leveraging MediaPipe landmark detection. Virtual Reality 27(4), 3125\u20133132 (2023)","journal-title":"Virtual Reality"},{"issue":"4","key":"1478_CR10","doi-asserted-by":"publisher","first-page":"601","DOI":"10.1109\/THMS.2016.2537749","volume":"46","author":"H Cecotti","year":"2016","unstructured":"Cecotti, H.: A multimodal gaze-controlled virtual keyboard. IEEE Trans. Human-Mach. Syst. 46(4), 601\u2013606 (2016). https:\/\/doi.org\/10.1109\/THMS.2016.2537749","journal-title":"IEEE Trans. Human-Mach. Syst."},{"issue":"7","key":"1478_CR11","doi-asserted-by":"publisher","first-page":"614","DOI":"10.12720\/jcm.14.7.614-621","volume":"14","author":"AF Klaib","year":"2019","unstructured":"Klaib, A.F., Alsrehin, N.O., Melhem, W.Y., Bashtawi, H.O.: IoT smart home using eye tracking and voice interfaces for elderly and special needs people. J. Commun. 14(7), 614\u2013621 (2019)","journal-title":"J. Commun."},{"key":"1478_CR12","doi-asserted-by":"publisher","first-page":"41952","DOI":"10.1109\/ACCESS.2022.3164075","volume":"10","author":"HO Edughele","year":"2022","unstructured":"Edughele, H.O., Zhang, Y., Muhammad-Sukki, F., Vien, Q.T., Morris-Cafiero, H., Agyeman, M.O.: Eye-tracking assistive technologies for individuals with amyotrophic lateral sclerosis. IEEE Access 10, 41952\u201341972 (2022). https:\/\/doi.org\/10.1109\/ACCESS.2022.3164075","journal-title":"IEEE Access"},{"key":"1478_CR13","doi-asserted-by":"publisher","first-page":"6","DOI":"10.1016\/j.visres.2016.09.002","volume":"128","author":"I Hooge","year":"2016","unstructured":"Hooge, I., Holmqvist, K., Nystr\u00f6m, M.: The pupil is faster than the corneal reflection (CR): are video based pupil-CR eye trackers suitable for studying detailed dynamics of eye movements? Vis. Res. 128, 6\u201318 (2016). https:\/\/doi.org\/10.1016\/j.visres.2016.09.002","journal-title":"Vis. Res."},{"key":"1478_CR14","doi-asserted-by":"publisher","first-page":"108944","DOI":"10.1016\/j.patcog.2022.108944","volume":"132","author":"J Liu","year":"2022","unstructured":"Liu, J., Chi, J., Yang, H., Yin, X.: In the eye of the beholder: a survey of gaze tracking techniques. Pattern Recognit. 132, 108944 (2022). https:\/\/doi.org\/10.1016\/j.patcog.2022.108944","journal-title":"Pattern Recognit."},{"issue":"9","key":"1478_CR15","first-page":"18","volume":"71","author":"A Sharma","year":"2013","unstructured":"Sharma, A., Abrol, P.: Eye gaze techniques for human computer interaction: a research survey. Int. J. Comput. Appl. 71(9), 18\u201325 (2013)","journal-title":"Int. J. Comput. Appl."},{"key":"1478_CR16","doi-asserted-by":"publisher","unstructured":"Cheng, Y., Wang, H., Bao, Y., Lu, F.: Appearance-based gaze estimation with deep learning: a review and benchmark. arXiv preprint arXiv:2104.12668 (2021). https:\/\/doi.org\/10.48550\/arXiv.2104.12668. Accessed 16 Dec 2023","DOI":"10.48550\/arXiv.2104.12668"},{"issue":"4","key":"1478_CR17","doi-asserted-by":"publisher","first-page":"409","DOI":"10.1007\/s10209-011-0256-6","volume":"11","author":"A Kr\u00f3lak","year":"2012","unstructured":"Kr\u00f3lak, A., Strumi\u0142\u0142o, P.: Eye-blink detection system for human\u2013computer interaction. Univ. Access Inf. Soc. 11(4), 409\u2013419 (2012)","journal-title":"Univ. Access Inf. Soc."},{"issue":"3","key":"1478_CR18","doi-asserted-by":"publisher","first-page":"915","DOI":"10.1109\/JBHI.2015.2412125","volume":"20","author":"ZP Bian","year":"2015","unstructured":"Bian, Z.P., Hou, J., Chau, L.P., Magnenat-Thalmann, N.: Facial position and expression-based human\u2013computer interface for persons with tetraplegia. IEEE J. Biomed. Health Inform. 20(3), 915\u2013924 (2015). https:\/\/doi.org\/10.1109\/JBHI.2015.2412125","journal-title":"IEEE J. Biomed. Health Inform."},{"issue":"2","key":"1478_CR19","doi-asserted-by":"publisher","first-page":"453","DOI":"10.1109\/TBME.2013.2280900","volume":"61","author":"Y Nam","year":"2013","unstructured":"Nam, Y., Koo, B., Cichocki, A., Choi, S.: GOM-Face: GKP, EOG, and EMG-based multimodal interface with application to humanoid robot control. IEEE Trans. Biomed. Eng. 61(2), 453\u2013462 (2013). https:\/\/doi.org\/10.1109\/TBME.2013.2280900","journal-title":"IEEE Trans. Biomed. Eng."},{"issue":"1","key":"1478_CR20","doi-asserted-by":"publisher","first-page":"6","DOI":"10.1109\/JBHI.2014.2328317","volume":"19","author":"V Mihajlovi\u0107","year":"2014","unstructured":"Mihajlovi\u0107, V., Grundlehner, B., Vullers, R., Penders, J.: Wearable, wireless EEG solutions in daily life applications: what are we missing? IEEE J. Biomed. Health Inform. 19(1), 6\u201321 (2014). https:\/\/doi.org\/10.1109\/JBHI.2014.2328317","journal-title":"IEEE J. Biomed. Health Inform."},{"key":"1478_CR21","doi-asserted-by":"publisher","first-page":"187","DOI":"10.1109\/RBME.2009.2035356","volume":"2","author":"J Mak","year":"2009","unstructured":"Mak, J., Wolpaw, J.: Clinical applications of brain-computer interfaces: current state and future prospects\u201d. IEEE Rev. Biomed. Eng. 2, 187\u2013199 (2009). https:\/\/doi.org\/10.1109\/RBME.2009.2035356","journal-title":"IEEE Rev. Biomed. Eng."},{"key":"1478_CR22","doi-asserted-by":"publisher","first-page":"205","DOI":"10.4018\/978-1-61350-098-9.ch015","volume-title":"Gaze Interaction and Applications of Eye Tracking: Advances in Assistive Technologies","author":"IS MacKenzie","year":"2012","unstructured":"MacKenzie, I.S.: Evaluating eye tracking systems for computer input\u201d. In: Ram, K. (ed.) Gaze Interaction and Applications of Eye Tracking: Advances in Assistive Technologies, pp. 205\u2013225. IGI Global, Hershey, PA, USA (2012). https:\/\/doi.org\/10.4018\/978-1-61350-098-9.ch015"},{"issue":"6","key":"1478_CR23","doi-asserted-by":"publisher","first-page":"979","DOI":"10.1109\/TNSRE.2013.2248748","volume":"21","author":"X Huo","year":"2013","unstructured":"Huo, X., Park, H., Kim, J., Ghovanloo, M.: A dual-mode human computer interface combining speech and tongue motion for people with severe disabilities. IEEE Trans. Neural Syst. Rehabil. Eng. 21(6), 979\u2013991 (2013). https:\/\/doi.org\/10.1109\/TNSRE.2013.2248748","journal-title":"IEEE Trans. Neural Syst. Rehabil. Eng."},{"issue":"1","key":"1478_CR24","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TNSRE.2002.1021581","volume":"10","author":"M Betke","year":"2002","unstructured":"Betke, M., Gips, J., Fleming, P.: The camera mouse: visual tracking of body features to provide computer access for people with severe disabilities. IEEE Trans. Neural Syst. Rehabil. Eng. 10(1), 1\u201310 (2002). https:\/\/doi.org\/10.1109\/TNSRE.2002.1021581","journal-title":"IEEE Trans. Neural Syst. Rehabil. Eng."},{"key":"1478_CR25","doi-asserted-by":"publisher","unstructured":"Tu, J., Huang, T., Tao, H.: Face as mouse through visual face tracking. In: The 2nd Canadian Conference on Computer and Robot Vision (CRV'05), pp. 339\u2013346 (2005). https:\/\/doi.org\/10.1109\/CRV.2005.39","DOI":"10.1109\/CRV.2005.39"},{"issue":"16","key":"1478_CR26","doi-asserted-by":"publisher","first-page":"3630","DOI":"10.3390\/s19163630","volume":"19","author":"RG Bozomitu","year":"2019","unstructured":"Bozomitu, R.G., P\u0103s\u0103ric\u0103, A., T\u0103rniceriu, D., Rotariu, C.: Development of an eye tracking-based human-computer interface for real-time applications. Sensors 19(16), 3630 (2019). https:\/\/doi.org\/10.3390\/s19163630","journal-title":"Sensors"},{"issue":"13","key":"1478_CR27","doi-asserted-by":"publisher","first-page":"18011","DOI":"10.1007\/s11042-022-12775-6","volume":"81","author":"D Bisen","year":"2022","unstructured":"Bisen, D., Shukla, R., Rajpoot, N., Maurya, P., Uttam, A.K.: Responsive human-computer interaction model based on recognition of facial landmarks using machine learning algorithms. Multimed. Tools Appl. 81(13), 18011\u201318031 (2022). https:\/\/doi.org\/10.1007\/s11042-022-12775-6","journal-title":"Multimed. Tools Appl."},{"key":"1478_CR28","doi-asserted-by":"publisher","DOI":"10.1155\/2017\/5739301","author":"X Zhang","year":"2017","unstructured":"Zhang, X., Liu, X., Yuan, S.M., Lin, S.F.: Eye tracking based control system for natural human-computer interaction. Comput. Intell. Neurosci. (2017). https:\/\/doi.org\/10.1155\/2017\/5739301","journal-title":"Comput. Intell. Neurosci."},{"key":"1478_CR29","unstructured":"GitHub, \u201cMasterLomaster\/bkb,\u201d 2015, https:\/\/github.com\/MastaLomaster\/bkb. Accessed 16 Dec 2023"},{"issue":"4","key":"1478_CR30","doi-asserted-by":"publisher","first-page":"2410","DOI":"10.11591\/eei.v12i4.5591","volume":"12","author":"RR Hossieny","year":"2023","unstructured":"Hossieny, R.R., Tantawi, M., Shedeed, H., Tolba, M.F.: Development of electrooculogram based human computer interface system using deep learning. Bull. Electr. Eng. Inform. 12(4), 2410\u20132420 (2023). https:\/\/doi.org\/10.11591\/eei.v12i4.5591","journal-title":"Bull. Electr. Eng. Inform."},{"issue":"1","key":"1478_CR31","doi-asserted-by":"publisher","first-page":"37","DOI":"10.1109\/TNSRE.2016.2542524","volume":"25","author":"KR Lee","year":"2016","unstructured":"Lee, K.R., Chang, W.D., Kim, S., Im, C.H.: Real-time \u201ceye-writing\u201d recognition using electrooculogram. IEEE Trans. Neural Syst. Rehabil. Eng. 25(1), 37\u201348 (2016). https:\/\/doi.org\/10.1109\/TNSRE.2016.2542524","journal-title":"IEEE Trans. Neural Syst. Rehabil. Eng."},{"issue":"10","key":"1478_CR32","doi-asserted-by":"publisher","first-page":"2404","DOI":"10.1016\/j.clinph.2021.07.012","volume":"132","author":"C Verbaarschot","year":"2021","unstructured":"Verbaarschot, C., Tump, D., Lutu, A., Borhanazad, M., Thielen, J., van den Broek, P., Farquhar, J., Weikamp, J., Raaphorst, J., Groothuis, J.T., Desain, P.: A visual brain-computer interface as communication aid for patients with amyotrophic lateral sclerosis. Clin. Neurophysiol. 132(10), 2404\u20132415 (2021). https:\/\/doi.org\/10.1016\/j.clinph.2021.07.012","journal-title":"Clin. Neurophysiol."},{"issue":"15","key":"1478_CR33","doi-asserted-by":"publisher","first-page":"19679","DOI":"10.1007\/s11042-017-5426-y","volume":"77","author":"C Zhang","year":"2018","unstructured":"Zhang, C., Yao, R., Cai, J.: Efficient eye typing with 9-direction gaze estimation. Multimed. Tools Appl. 77(15), 19679\u201319696 (2018). https:\/\/doi.org\/10.1007\/s11042-017-5426-y","journal-title":"Multimed. Tools Appl."},{"key":"1478_CR34","doi-asserted-by":"publisher","DOI":"10.1155\/2021\/8213946","author":"J Huang","year":"2021","unstructured":"Huang, J., Zhang, Z., Xie, G., He, H.: Real-time precise human-computer interaction system based on gaze estimation and tracking. Wirel. Commun. Mob. Comput. (2021). https:\/\/doi.org\/10.1155\/2021\/8213946","journal-title":"Wirel. Commun. Mob. Comput."},{"key":"1478_CR35","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-022-13085-7","author":"K Donuk","year":"2022","unstructured":"Donuk, K., Ari, A., Hanbay, D.: A CNN based real-time eye tracker for web mining applications. Multimed. Tools Appl. (2022). https:\/\/doi.org\/10.1007\/s11042-022-13085-7","journal-title":"Multimed. Tools Appl."},{"issue":"8","key":"1478_CR36","doi-asserted-by":"publisher","first-page":"4138","DOI":"10.3390\/s23084138","volume":"23","author":"MF Ansari","year":"2023","unstructured":"Ansari, M.F., Kasprowski, P., Peer, P.: Person-specific gaze estimation from low-quality webcam images. Sensors 23(8), 4138 (2023). https:\/\/doi.org\/10.3390\/s23084138","journal-title":"Sensors"},{"key":"1478_CR37","doi-asserted-by":"crossref","unstructured":"LRD, M., Mukhopadhyay, A., Biswas, P.: Distraction detection in automotive environment using appearance-based gaze estimation. In: 27th International Conference on Intelligent User Interfaces, pp. 38\u201341(2022)","DOI":"10.1145\/3490100.3516463"},{"issue":"4","key":"1478_CR38","doi-asserted-by":"publisher","first-page":"319","DOI":"10.3233\/TAD-200292","volume":"33","author":"DV Jeevithashree","year":"2021","unstructured":"Jeevithashree, D.V., Jain, P., Mukhopadhyay, A., Saluja, K.P.S., Biswas, P.: Eye gaze controlled adaptive virtual keyboard for users with SSMI. Technol. Disabil. 33(4), 319\u2013338 (2021)","journal-title":"Technol. Disabil."},{"key":"1478_CR39","unstructured":"MS Windows NT kernel description, https:\/\/google.github.io\/mediapipe\/solutions\/face_mesh.html. Accessed 10 June 2023"},{"key":"1478_CR40","doi-asserted-by":"publisher","DOI":"10.1155\/2021\/8133076","author":"S Shriram","year":"2021","unstructured":"Shriram, S., Nagaraj, B., Jaya, J., Shankar, S., Ajay, P.: Deep learning-based real-time AI virtual mouse system using computer vision to avoid COVID-19 spread. J. Healthc. Eng. (2021). https:\/\/doi.org\/10.1155\/2021\/8133076","journal-title":"J. Healthc. Eng."},{"key":"1478_CR41","doi-asserted-by":"publisher","unstructured":"Ablavatski, A., Vakunov, A., Grishchenko, I., Raveendran, K., Zhdanovich, M.: Real-time pupil tracking from monocular video for digital puppetry. arXiv preprint arXiv:2006.11341 (2020). https:\/\/doi.org\/10.48550\/arXiv.2006.11341. Accessed 16 Dec 2023","DOI":"10.48550\/arXiv.2006.11341"},{"key":"1478_CR42","unstructured":"Kartynnik, Y., Ablavatski, A., Grishchenko, I., Grundmann, M.: Real-time facial surface geometry from monocular video on mobile GPUs. arXiv preprint arXiv:1907.06724 (2019). Accessed 16 Dec 2023"},{"key":"1478_CR43","doi-asserted-by":"crossref","unstructured":"Caporusso, N., Sanders, G., Thaman, B., Hall, E.: An eye-tracking solution using consumer grade webcams for potential concussion diagnosis and evaluation. In: 2023 46th MIPRO ICT and Electronics Convention (MIPRO), pp. 67\u201372 (2023)","DOI":"10.23919\/MIPRO57284.2023.10159830"},{"issue":"5","key":"1478_CR44","doi-asserted-by":"publisher","first-page":"522","DOI":"10.1111\/cxo.12583","volume":"100","author":"JP Bergmanson","year":"2017","unstructured":"Bergmanson, J.P., Martinez, J.G.: Size does matter: what is the corneo-limbal diameter? Clin. Exp. Optom. 100(5), 522\u2013528 (2017). https:\/\/doi.org\/10.1111\/cxo.12583","journal-title":"Clin. Exp. Optom."},{"key":"1478_CR45","volume-title":"Fundamentals of optics","author":"FA Jenkins","year":"2002","unstructured":"Jenkins, F.A., White, H.E.: Fundamentals of optics. McGraw-Hill Companies, New York (2002)"},{"issue":"4","key":"1478_CR46","doi-asserted-by":"publisher","first-page":"04023014","DOI":"10.1061\/JCCEE5.CPENG-5060","volume":"37","author":"S Chang","year":"2023","unstructured":"Chang, S., Siu, M.F.F., Li, H.: Development of a fuzzy logic controller for autonomous navigation of building inspection robots in unknown environments. J. Comput. Civ. Eng. 37(4), 04023014 (2023). https:\/\/doi.org\/10.1061\/JCCEE5.CPENG-5060","journal-title":"J. Comput. Civ. Eng."},{"key":"1478_CR47","doi-asserted-by":"crossref","unstructured":"Soukoreff, R.W., MacKenzie, I.S.: Metrics for text entry research: an evaluation of MSD and KSPC, and a new unified error metric. In: Proceedings of the SIGCHI conference on Human factors in computing systems, pp. 113\u2013120 (2003)","DOI":"10.1145\/642611.642632"},{"issue":"9","key":"1478_CR48","doi-asserted-by":"publisher","first-page":"3151","DOI":"10.3390\/s18093151","volume":"18","author":"A Kar","year":"2018","unstructured":"Kar, A., Corcoran, P.: Performance evaluation strategies for eye gaze estimation systems with quantitative metrics and visualizations. Sensors 18(9), 3151 (2018). https:\/\/doi.org\/10.3390\/s18093151","journal-title":"Sensors"},{"issue":"3","key":"1478_CR49","first-page":"114","volume":"4","author":"A Bangor","year":"2009","unstructured":"Bangor, A., Kortum, P., Miller, J.: Determining what individual SUS scores mean: adding an adjective rating scale. J. Usability Stud. 4(3), 114\u2013123 (2009)","journal-title":"J. Usability Stud."}],"container-title":["Journal of Real-Time Image Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-024-01478-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11554-024-01478-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-024-01478-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,19]],"date-time":"2024-11-19T12:17:49Z","timestamp":1732018669000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11554-024-01478-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,20]]},"references-count":49,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2024,6]]}},"alternative-id":["1478"],"URL":"https:\/\/doi.org\/10.1007\/s11554-024-01478-w","relation":{},"ISSN":["1861-8200","1861-8219"],"issn-type":[{"value":"1861-8200","type":"print"},{"value":"1861-8219","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,5,20]]},"assertion":[{"value":"6 February 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 May 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 May 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"To maintain the integrity and objectivity of our work, we would like to state that the authors have no conflicts of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Ethical approval was obtained from Thapar Institute of Engineering & Technology, with approval number TIET\/EC\/2023-13.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"Before participating in the study, all participants were provided with detailed information about the research objectives, procedures, and benefits. Informed verbal consent was obtained from each participant, indicating their voluntary agreement to participate in the study. They were also assured of their data\u2019s confidentiality and personal information.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"Participants were informed that the research findings during the study would be used for publication in journals. Verbal consent for the publication of study outcomes was obtained from all participants. They were assured that any information disclosed in the manuscript would maintain their anonymity and confidentiality.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to publish"}}],"article-number":"97"}}