{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,28]],"date-time":"2026-03-28T22:25:41Z","timestamp":1774736741528,"version":"3.50.1"},"reference-count":43,"publisher":"Springer Science and Business Media LLC","issue":"7","license":[{"start":{"date-parts":[[2025,5,19]],"date-time":"2025-05-19T00:00:00Z","timestamp":1747612800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,5,19]],"date-time":"2025-05-19T00:00:00Z","timestamp":1747612800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2025,7]]},"DOI":"10.1007\/s40747-025-01897-7","type":"journal-article","created":{"date-parts":[[2025,5,19]],"date-time":"2025-05-19T08:41:15Z","timestamp":1747644075000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["AI-enabled driver assistance: monitoring head and gaze movements for enhanced safety"],"prefix":"10.1007","volume":"11","author":[{"given":"Sayyed","family":"Mudassar Shah","sequence":"first","affiliation":[]},{"given":"Gan","family":"Zengkang","sequence":"additional","affiliation":[]},{"given":"Zhaoyun","family":"Sun","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4761-0346","authenticated-orcid":false,"given":"Tariq","family":"Hussain","sequence":"additional","affiliation":[]},{"given":"Khalid","family":"Zaman","sequence":"additional","affiliation":[]},{"given":"Abdullah","family":"Alwabli","sequence":"additional","affiliation":[]},{"given":"Amar Y.","family":"Jaffar","sequence":"additional","affiliation":[]},{"given":"Farman","family":"Ali","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,19]]},"reference":[{"key":"1897_CR1","doi-asserted-by":"crossref","unstructured":"Alsajri A (2023) A review on machine learning strategies for Real-World engineering applications. Babylon J Mach Learn, vol. pp. 1\u20136, 01\/15 2023.","DOI":"10.58496\/BJML\/2023\/001"},{"key":"1897_CR2","doi-asserted-by":"crossref","unstructured":"Chong E, Wang Y, Ruiz N, Rehg JM (2020) Detecting attended visual targets in video, in Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 5396\u20135406","DOI":"10.1109\/CVPR42600.2020.00544"},{"key":"1897_CR3","doi-asserted-by":"crossref","unstructured":"Sugano Y, Zhang X, Bulling A (2016) Aggregaze: Collective estimation of audience attention on public displays, in Proceedings of the 29th Annual Symposium on User Interface Software and Technology, pp. 821\u2013831","DOI":"10.1145\/2984511.2984536"},{"key":"1897_CR4","doi-asserted-by":"crossref","unstructured":"Kodama Y, Kawanishi Y, Hirayama T, Deguchi D, Ide I, Murase H et al (2018) Localizing the gaze target of a crowd of people, in Computer Vision\u2013ACCV 2018 Workshops: 14th Asian Conference on Computer Vision, Perth, Australia, December 2\u20136, Revised Selected Papers 14, 2019, pp. 15\u201330.","DOI":"10.1007\/978-3-030-21074-8_2"},{"key":"1897_CR5","doi-asserted-by":"crossref","unstructured":"Bermejo C, Chatzopoulos D, Hui P (2020) Eyeshopper: Estimating shoppers\u2019 gaze using cctv cameras, in Proceedings of the 28th ACM International Conference on Multimedia, pp. 2765\u20132774","DOI":"10.1145\/3394171.3413683"},{"key":"1897_CR6","doi-asserted-by":"crossref","unstructured":"Zaman K, Zhaoyun S, Shah B, Hussain T, Shah SM, Ali F, Khan US (2023) A novel driver emotion recognition system based on deep ensemble classification. Complex Intell Syst 9(6):6927\u20136952","DOI":"10.1007\/s40747-023-01100-9"},{"key":"1897_CR7","doi-asserted-by":"publisher","first-page":"546","DOI":"10.1109\/THMS.2015.2504083","volume":"46","author":"R Zheng","year":"2015","unstructured":"Zheng R, Nakano K, Ishiko H, Hagita K, Kihira M, Yokozeki T (2015) Eye-gaze tracking analysis of driver behavior while interacting with navigation systems in an urban area. IEEE Trans Human-Machine Syst 46:546\u2013556","journal-title":"IEEE Trans Human-Machine Syst"},{"key":"1897_CR8","doi-asserted-by":"crossref","unstructured":"Zaman K, Sun Z, Shah SM, Shoaib M, Pei L, Hussain A (2022) Driver Emotions Recognition Based on Improved Faster R-CNN and Neural Architectural Search Network, Symmetry, vol. 14, p. 687","DOI":"10.3390\/sym14040687"},{"key":"1897_CR9","unstructured":"Cheng Y, Wang H, Bao Y, Lu F (2021) Appearance-based gaze estimation with deep learning: A review and benchmark, arXiv preprint arXiv:2104.12668,"},{"key":"1897_CR10","doi-asserted-by":"publisher","first-page":"16495","DOI":"10.1109\/ACCESS.2017.2735633","volume":"5","author":"A Kar","year":"2017","unstructured":"Kar A, Corcoran P (2017) A review and analysis of eye-gaze Estimation systems, algorithms and performance evaluation methods in consumer platforms. IEEE Access 5:16495\u201316519","journal-title":"IEEE Access"},{"key":"1897_CR11","doi-asserted-by":"publisher","first-page":"142581","DOI":"10.1109\/ACCESS.2020.3013540","volume":"8","author":"AA Akinyelu","year":"2020","unstructured":"Akinyelu AA, Blignaut P (2020) Convolutional neural network-based methods for eye gaze estimation: A survey. IEEE Access 8:142581\u2013142605","journal-title":"IEEE Access"},{"key":"1897_CR12","doi-asserted-by":"crossref","unstructured":"Kellnhofer P, Recasens A, Stent S, Matusik W, Torralba A (2019) Gaze360: Physically unconstrained gaze estimation in the wild, in Proceedings of the IEEE\/CVF international conference on computer vision, pp. 6912\u20136921","DOI":"10.1109\/ICCV.2019.00701"},{"key":"1897_CR13","doi-asserted-by":"publisher","first-page":"3739","DOI":"10.3390\/s20133739","volume":"20","author":"D Cazzato","year":"2020","unstructured":"Cazzato D, Leo M, Distante C, Voos H (2020) When i look into your eyes: A survey on computer vision contributions for human gaze Estimation and tracking. Sensors 20:3739","journal-title":"Sensors"},{"key":"1897_CR14","doi-asserted-by":"crossref","unstructured":"Guo H, Hu Z, Liu J (2022) MGTR: End-to-End Mutual Gaze Detection with Transformer, in Proceedings of the Asian Conference on Computer Vision, pp. 1590\u20131605","DOI":"10.1007\/978-3-031-26316-3_22"},{"key":"1897_CR15","doi-asserted-by":"crossref","unstructured":"Shah SM, Sun Z, Zaman K, Hussain A, Shoaib M, Pei L (2022) A Driver Gaze Estimation Method Based on Deep Learning, Sensors, vol. 22, p. 3959","DOI":"10.3390\/s22103959"},{"key":"1897_CR16","doi-asserted-by":"publisher","first-page":"631","DOI":"10.1016\/S0169-8141(98)00068-7","volume":"24","author":"JH Goldberg","year":"1999","unstructured":"Goldberg JH, Kotval XP (1999) Computer interface evaluation using eye movements: methods and constructs. Int J Ind Ergon 24:631\u2013645","journal-title":"Int J Ind Ergon"},{"key":"1897_CR17","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T et al (2010),., An image is worth 16x16 words: Transformers for image recognition at scale, arXiv preprint arXiv:11929, 2020."},{"key":"1897_CR18","doi-asserted-by":"publisher","first-page":"233","DOI":"10.1007\/s11633-020-1252-1","volume":"18","author":"S De Silva","year":"2021","unstructured":"De Silva S, Dayarathna S, Ariyarathne G, Meedeniya D, Jayarathna S, Michalek AM (2021) Computational decision support system for ADHD identification. Int J Autom Comput 18:233\u2013255","journal-title":"Int J Autom Comput"},{"key":"1897_CR19","doi-asserted-by":"publisher","first-page":"75","DOI":"10.1016\/j.procs.2014.11.012","volume":"39","author":"BB Velichkovsky","year":"2014","unstructured":"Velichkovsky BB, Rumyantsev MA, Morozov MA (2014) New solution to the Midas touch problem: identification of visual commands via extraction of focal fixations. Procedia Comput Sci 39:75\u201382","journal-title":"Procedia Comput Sci"},{"key":"1897_CR20","doi-asserted-by":"publisher","first-page":"41","DOI":"10.20870\/IJVR.2006.5.3.2697","volume":"5","author":"Y-M Kwon","year":"2006","unstructured":"Kwon Y-M, Jeon K-W, Ki J, Shahab QM, Jo S, Kim S-K (2006) 3D gaze Estimation and interaction to stereo display. Int J Virtual Real 5:41\u201345","journal-title":"Int J Virtual Real"},{"key":"1897_CR21","unstructured":"Krishna GS, Supriya K, Vardhan J (2022) Vision transformers and YoloV5 based driver drowsiness detection framework, arXiv preprint arXiv:2209.01401,"},{"key":"1897_CR22","doi-asserted-by":"crossref","unstructured":"Rangesh A, Zhang B, Trivedi MM (2020) Driver gaze estimation in the real world: Overcoming the eyeglass challenge, in 2020 IEEE Intelligent Vehicles Symposium (IV), pp. 1054\u20131059","DOI":"10.1109\/IV47402.2020.9304573"},{"key":"1897_CR23","unstructured":"Pan J, Ferrer CC, McGuinness K, O\u2019Connor NE, Torres J, Sayrol E et al (2017) Salgan: Visual saliency prediction with generative adversarial networks, arXiv preprint arXiv:1701.01081,"},{"key":"1897_CR24","unstructured":"Taigman Y, Polyak A, Wolf L (2016) Unsupervised cross-domain image generation, arXiv preprint arXiv:1611.02200,"},{"key":"1897_CR25","doi-asserted-by":"crossref","unstructured":"Yu Z, Huang X, Zhang X, Shen H, Li Q, Deng W et al (2020) A multi-modal approach for driver gaze prediction to remove identity bias, in Proceedings of the International Conference on Multimodal Interaction, 2020, pp. 768\u2013776.","DOI":"10.1145\/3382507.3417961"},{"key":"1897_CR26","doi-asserted-by":"publisher","first-page":"8715","DOI":"10.1109\/TITS.2021.3085492","volume":"23","author":"JA Abbasi","year":"2021","unstructured":"Abbasi JA, Mullins D, Ringelstein N, Reilhac P, Jones E, Glavin M (2021) An analysis of driver gaze behaviour at roundabouts. IEEE Trans Intell Transp Syst 23:8715\u20138724","journal-title":"IEEE Trans Intell Transp Syst"},{"key":"1897_CR27","doi-asserted-by":"publisher","first-page":"12326","DOI":"10.1109\/JSEN.2020.2999625","volume":"20","author":"J Sun","year":"2020","unstructured":"Sun J, Lu S (2020) An improved single shot multibox for video-rate head pose prediction. IEEE Sens J 20:12326\u201312333","journal-title":"IEEE Sens J"},{"key":"1897_CR28","doi-asserted-by":"crossref","unstructured":"Ledezma A, Zamora V, Sipele \u00d3, Sesmero MP, Sanchis A (2021) Implementing a gaze tracking algorithm for improving advanced driver assistance systems, Electronics, vol. 10, p. 1480","DOI":"10.3390\/electronics10121480"},{"key":"1897_CR29","doi-asserted-by":"publisher","first-page":"1800","DOI":"10.1109\/TIE.2021.3057033","volume":"69","author":"Z Hu","year":"2021","unstructured":"Hu Z, Lv C, Hang P, Huang C, Xing Y (2021) Data-driven Estimation of driver attention using calibration-free eye gaze and scene features. IEEE Trans Industr Electron 69:1800\u20131808","journal-title":"IEEE Trans Industr Electron"},{"key":"1897_CR30","doi-asserted-by":"crossref","unstructured":"Tang F, Yang F, Tian X (2023) Long-Distance Person Detection Based on YOLOv7, Electronics, vol. 12, p. 1502","DOI":"10.3390\/electronics12061502"},{"key":"1897_CR31","unstructured":"Park H, Jain E, Sheikh Y (2012) 3d social saliency from head-mounted cameras. Adv Neural Inf Process Syst, 25"},{"key":"1897_CR32","doi-asserted-by":"crossref","unstructured":"Kumar M, Paepcke A, Winograd T (2007) Eyepoint: practical pointing and selection using gaze and keyboard, in Proceedings of the SIGCHI conference on Human factors in computing systems, pp. 421\u2013430","DOI":"10.1145\/1240624.1240692"},{"key":"1897_CR33","doi-asserted-by":"crossref","unstructured":"Li J, Liu C, Lu X, Wu B (2022) CME-YOLOv5: An efficient object detection network for densely spaced fish and small targets, Water, vol. 14, p. 2412","DOI":"10.3390\/w14152412"},{"key":"1897_CR34","doi-asserted-by":"publisher","first-page":"5540","DOI":"10.3390\/s19245540","volume":"19","author":"MQ Khan","year":"2019","unstructured":"Khan MQ, Lee S (2019) Gaze and eye tracking: techniques and applications in ADAS. Sensors 19:5540","journal-title":"Sensors"},{"key":"1897_CR35","doi-asserted-by":"publisher","unstructured":"Sun W, Zhang J, Cai C, Lin X, Pan Z, Tariq H (2025) Efficient classroom behavior detection through end-to end multiscale feature fusion. In: Proceedings of the 19th ACM SIGGRAPH international conference on virtual-reality continuum and its applications in industry (VRCAI\u201924). Association for Computing Machinery, New York, Article 13, pp 1\u20137. https:\/\/doi.org\/10.1145\/3703619.3706046","DOI":"10.1145\/3703619.3706046"},{"key":"1897_CR36","doi-asserted-by":"crossref","unstructured":"Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: Unified, real-time object detection, in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 779\u2013788","DOI":"10.1109\/CVPR.2016.91"},{"key":"1897_CR37","doi-asserted-by":"crossref","unstructured":"Woo S, Park J, Lee J-Y, Kweon IS (2018) Cbam: Convolutional block attention module, in Proceedings of the European conference on computer vision (ECCV), pp. 3\u201319","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"1897_CR38","doi-asserted-by":"crossref","unstructured":"Sheng Z, Chen H, Qi Z (2023) CBAM-based Method in YOLOv7 for Detecting Defective Vacuum Glass Tubes, in Proceedings of the 2nd Asia Conference on Algorithms, Computing and Machine Learning, 2023, pp. 413\u2013418","DOI":"10.1145\/3590003.3590079"},{"key":"1897_CR39","doi-asserted-by":"publisher","first-page":"6991","DOI":"10.1038\/s41598-022-11173-0","volume":"12","author":"T Debnath","year":"2022","unstructured":"Debnath T, Reza MM, Rahman A, Beheshti A, Band SS, Alinejad-Rokny H (2022) Four-layer ConvNet to facial emotion recognition with minimal epochs and the significance of data diversity. Sci Rep 12:6991","journal-title":"Sci Rep"},{"key":"1897_CR40","doi-asserted-by":"crossref","unstructured":"Zhang X, Sugano Y, Fritz M, Bulling A (2017) MPIIGaze: real-world dataset and deep appearance-based gaze estimation. IEEE Trans Pattern Anal Mach Intell 41(1):162. arXiv:1711.09017","DOI":"10.1109\/TPAMI.2017.2778103"},{"key":"1897_CR41","doi-asserted-by":"crossref","unstructured":"Zhang X, Sugano Y, Fritz M, Bulling A (2015) Appearance-based gaze estimation in the wild. In: Proceedings of the IEEE conference on computer vision and pattern recognition  (CVPR), 2015. arXiv:1504.02863, Project Page","DOI":"10.1109\/CVPR.2015.7299081"},{"key":"1897_CR42","doi-asserted-by":"publisher","unstructured":"Ribeiro RF, Costa PD (2019) Driver gaze zone dataset with depth data. In: 2019 14th IEEE international conference on automatic face & gesture recognition (FG 2019), Lille, France, pp 1\u20135. https:\/\/doi.org\/10.1109\/FG.2019.8756592","DOI":"10.1109\/FG.2019.8756592"},{"key":"1897_CR43","doi-asserted-by":"crossref","unstructured":"Funes Mora KA, Monay F, Odobez JM (2014) Eyediap: a database for the development and evaluation of gaze estimation algorithms from rgb and rgb-d cameras. In: Proceedings of the symposium on eye tracking research and applications, pp 255\u2013258. ACM","DOI":"10.1145\/2578153.2578190"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-025-01897-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-025-01897-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-025-01897-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T11:08:56Z","timestamp":1750331336000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-025-01897-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,19]]},"references-count":43,"journal-issue":{"issue":"7","published-print":{"date-parts":[[2025,7]]}},"alternative-id":["1897"],"URL":"https:\/\/doi.org\/10.1007\/s40747-025-01897-7","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,5,19]]},"assertion":[{"value":"22 June 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 March 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 May 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"We confirm that relevant guidelines and regulations are carried out in all methods.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"The authors declare no competing interests.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"297"}}