{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,27]],"date-time":"2025-07-27T07:26:09Z","timestamp":1753601169546,"version":"3.40.4"},"reference-count":153,"publisher":"Springer Science and Business Media LLC","issue":"12","license":[{"start":{"date-parts":[[2024,5,18]],"date-time":"2024-05-18T00:00:00Z","timestamp":1715990400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,5,18]],"date-time":"2024-05-18T00:00:00Z","timestamp":1715990400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-024-19368-5","type":"journal-article","created":{"date-parts":[[2024,5,18]],"date-time":"2024-05-18T06:01:46Z","timestamp":1716012106000},"page":"10969-11000","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Recent advancements in driver\u2019s attention prediction"],"prefix":"10.1007","volume":"84","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6825-5171","authenticated-orcid":false,"given":"Morteza","family":"Moradi","sequence":"first","affiliation":[]},{"given":"Simone","family":"Palazzo","sequence":"additional","affiliation":[]},{"given":"Francesco","family":"Rundo","sequence":"additional","affiliation":[]},{"given":"Concetto","family":"Spampinato","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,5,18]]},"reference":[{"key":"19368_CR1","doi-asserted-by":"crossref","unstructured":"Abla\u00dfmeier M, Poitschke T, Wallhoff F, et\u00a0al (2007) Eye gaze studies comparing head-up and head-down displays in vehicles. In: 2007 IEEE International Conference on Multimedia and Expo, IEEE, pp 2250\u20132252","DOI":"10.1109\/ICME.2007.4285134"},{"key":"19368_CR2","unstructured":"Aksoy E, Yazici A, Kasap M (2020) See, attend and brake: An attention-based saliency map prediction model for end-to-end driving. CoRR abs\/2002.11020. https:\/\/arxiv.org\/abs\/2002.11020, 2002.11020"},{"key":"19368_CR3","doi-asserted-by":"crossref","unstructured":"Almahasneh H, Chooi WT, Kamel N et al (2014) Deep in thought while driving: An eeg study on drivers\u2019 cognitive distraction. Transportation research part F: traffic psychology and behaviour 26:218\u2013226","DOI":"10.1016\/j.trf.2014.08.001"},{"key":"19368_CR4","unstructured":"Ameyoe A, Mars F, Chevrel P et al (2015) Estimation of driver distraction using the prediction error of a cybernetic driver model. DSC, Simulation Design and Architecture, Germany"},{"key":"19368_CR5","doi-asserted-by":"crossref","unstructured":"Araluce J, Bergasa LM, Oca\u00f1a M, et\u00a0al (2022) Aragan: A driver attention estimation model based on conditional generative adversarial network. In: 2022 IEEE Intelligent Vehicles Symposium (IV), IEEE, pp 1066\u20131072","DOI":"10.1109\/IV51971.2022.9827175"},{"key":"19368_CR6","doi-asserted-by":"crossref","unstructured":"Azman A, Meng Q, Edirisinghe E (2010) Non intrusive physiological measurement for driver cognitive distraction detection: Eye and mouth movements. In: 2010 3rd International Conference on Advanced Computer Theory and Engineering (ICACTE), IEEE, pp V3\u2013595","DOI":"10.1109\/ICACTE.2010.5579547"},{"key":"19368_CR7","doi-asserted-by":"crossref","unstructured":"Bach KM, J\u00e6ger MG, Skov MB, et\u00a0al (2009) Interacting with in-vehicle systems: understanding, measuring, and evaluating attention. In: Proceedings of the 2009 British Computer Society Conference on Human-Computer Interaction, BCS-HCI 2009, Cambridge, United Kingdom, 1-5 September 2009. ACM, pp 453\u2013462, https:\/\/dl.acm.org\/citation.cfm?id=1671070","DOI":"10.14236\/ewic\/HCI2009.57"},{"key":"19368_CR8","doi-asserted-by":"crossref","unstructured":"Baee S, Pakdamanian E, Kim I, et\u00a0al (2021) Medirl: Predicting the visual attention of drivers via maximum entropy deep inverse reinforcement learning. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 13178\u201313188","DOI":"10.1109\/ICCV48922.2021.01293"},{"key":"19368_CR9","doi-asserted-by":"crossref","unstructured":"Bao W, Yu Q, Kong Y (2021) Drive: Deep reinforced accident anticipation with visual explanation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 7619\u20137628","DOI":"10.1109\/ICCV48922.2021.00752"},{"key":"19368_CR10","unstructured":"Bazzani L, Larochelle H, Torresani L (2016) Recurrent mixture density network for spatiotemporal visual attention. arXiv preprint arXiv:1603.08199"},{"issue":"1","key":"19368_CR11","doi-asserted-by":"crossref","first-page":"55","DOI":"10.1109\/TIP.2012.2210727","volume":"22","author":"A Borji","year":"2012","unstructured":"Borji A, Sihite DN, Itti L (2012) Quantitative analysis of human-model agreement in visual saliency modeling: A comparative study. IEEE Trans Image Process 22(1):55\u201369","journal-title":"IEEE Trans Image Process"},{"key":"19368_CR12","doi-asserted-by":"crossref","unstructured":"Borji A, Tavakoli HR, Sihite DN, et\u00a0al (2013) Analysis of scores, datasets, and models in visual saliency prediction. In: Proceedings of the IEEE international conference on computer vision, pp 921\u2013928","DOI":"10.1109\/ICCV.2013.118"},{"key":"19368_CR13","doi-asserted-by":"crossref","first-page":"582","DOI":"10.1016\/j.neubiorev.2021.02.002","volume":"125","author":"N Brunkhorst-Kanaan","year":"2021","unstructured":"Brunkhorst-Kanaan N, Libutzki B, Reif A et al (2021) Adhd and accidents over the life span-a systematic review. Neurosci Biobehav Rev 125:582\u2013591","journal-title":"Neurosci Biobehav Rev"},{"issue":"3","key":"19368_CR14","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1109\/TPAMI.2018.2815601","volume":"41","author":"Z Bylinskii","year":"2019","unstructured":"Bylinskii Z, Judd T, Oliva A et al (2019) What do different evaluation metrics tell us about saliency models? IEEE Transactions on Pattern Analysis and Machine Intelligence 41(3):740\u2013757. https:\/\/doi.org\/10.1109\/TPAMI.2018.2815601","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"19368_CR15","doi-asserted-by":"crossref","first-page":"428","DOI":"10.1016\/j.neucom.2020.03.111","volume":"407","author":"J Cai","year":"2020","unstructured":"Cai J, Hu J, Tang X et al (2020) Deep historical long short-term memory network for action recognition. Neurocomputing 407:428\u2013438","journal-title":"Neurocomputing"},{"key":"19368_CR16","doi-asserted-by":"crossref","unstructured":"Chaabane M, Trabelsi A, Blanchard N, et\u00a0al (2020) Looking ahead: Anticipating pedestrians crossing with future frames prediction. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp 2297\u20132306","DOI":"10.1109\/WACV45572.2020.9093426"},{"key":"19368_CR17","doi-asserted-by":"crossref","unstructured":"Chang Z, Zhang X, Wang S, et\u00a0al (2022) Stam: A spatiotemporal attention based memory for video prediction. IEEE Transactions on Multimedia","DOI":"10.1109\/TMM.2022.3146721"},{"key":"19368_CR18","unstructured":"Chen L, Papandreou G, Schroff F, et\u00a0al (2017) Rethinking atrous convolution for semantic image segmentation. CoRR abs\/1706.05587. https:\/\/arxiv.org\/abs\/1706.05587, 1706.05587"},{"key":"19368_CR19","doi-asserted-by":"publisher","unstructured":"Chen Y, Nan Z (2022) Inspired by norbert wiener: Feedback loop network learning incremental knowledge for driver attention prediction and beyond. CoRR abs\/2212.02096. https:\/\/doi.org\/10.48550\/arXiv.2212.02096,2212.02096ff","DOI":"10.48550\/arXiv.2212.02096"},{"key":"19368_CR20","doi-asserted-by":"publisher","unstructured":"Chen Y, Wang J, Li J, et\u00a0al (2018) Lidar-video driving dataset: Learning driving policies effectively. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 5870\u20135878, https:\/\/doi.org\/10.1109\/CVPR.2018.00615","DOI":"10.1109\/CVPR.2018.00615"},{"key":"19368_CR21","doi-asserted-by":"crossref","unstructured":"Cheng F, Bertasius G (2022) Tallformer: Temporal action localization with a long-memory transformer. In: Computer Vision\u2013ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23\u201327, 2022, Proceedings, Part XXXIV, Springer, pp 503\u2013521","DOI":"10.1007\/978-3-031-19830-4_29"},{"key":"19368_CR22","doi-asserted-by":"crossref","unstructured":"Choi J, Chun D, Kim H, et\u00a0al (2019) Gaussian yolov3: An accurate and fast object detector using localization uncertainty for autonomous driving. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 502\u2013511","DOI":"10.1109\/ICCV.2019.00059"},{"key":"19368_CR23","first-page":"205566831877399","volume":"5","author":"M Cognolato","year":"2018","unstructured":"Cognolato M, Atzori M, M\u00fcller H (2018) Head-mounted eye gaze tracking devices: An overview of modern devices and recent advances. J Rehabil Assist Technol Eng 5:2055668318773991","journal-title":"J Rehabil Assist Technol Eng"},{"key":"19368_CR24","doi-asserted-by":"crossref","unstructured":"Cornia M, Baraldi L, Serra G, et\u00a0al (2016) A deep multi-level network for saliency prediction. In: 2016 23rd International Conference on Pattern Recognition (ICPR), IEEE, pp 3488\u20133493","DOI":"10.1109\/ICPR.2016.7900174"},{"issue":"10","key":"19368_CR25","doi-asserted-by":"crossref","first-page":"5142","DOI":"10.1109\/TIP.2018.2851672","volume":"27","author":"M Cornia","year":"2018","unstructured":"Cornia M, Baraldi L, Serra G et al (2018) Predicting human eye fixations via an lstm-based saliency attentive model. IEEE Trans Image Process 27(10):5142\u20135154","journal-title":"IEEE Trans Image Process"},{"issue":"1","key":"19368_CR26","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1186\/s12544-019-0384-2","volume":"11","author":"T Cvahte Ojster\u0161ek","year":"2019","unstructured":"Cvahte Ojster\u0161ek T, Topol\u0161ek D (2019) Influence of drivers\u2019 visual and cognitive attention on their perception of changes in the traffic environment. Eur Transp Res Rev 11(1):1\u20139","journal-title":"Eur Transp Res Rev"},{"key":"19368_CR27","doi-asserted-by":"crossref","unstructured":"Dai R, Minciullo L, Garattoni L, et\u00a0al (2019) Self-attention temporal convolutional network for long-term daily living activity detection. In: 2019 16th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), IEEE, pp 1\u20137","DOI":"10.1109\/AVSS.2019.8909841"},{"key":"19368_CR28","doi-asserted-by":"crossref","unstructured":"Deng J, Dong W, Socher R, et\u00a0al (2009) Imagenet: A large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition, Ieee, pp 248\u2013255","DOI":"10.1109\/CVPR.2009.5206848"},{"issue":"5","key":"19368_CR29","doi-asserted-by":"publisher","first-page":"2146","DOI":"10.1109\/TITS.2019.2915540","volume":"21","author":"T Deng","year":"2020","unstructured":"Deng T, Yan H, Qin L et al (2020) How do drivers allocate their potential attention? driving fixation prediction via convolutional neural networks. IEEE Trans Intell Transp Syst 21(5):2146\u20132154. https:\/\/doi.org\/10.1109\/TITS.2019.2915540","journal-title":"IEEE Trans Intell Transp Syst"},{"key":"19368_CR30","doi-asserted-by":"crossref","unstructured":"Deng T, Yan F, Yan H (2021) Driving video fixation prediction model via spatio-temporal networks and attention gates. In: 2021 IEEE International Conference on Multimedia and Expo (ICME), IEEE, pp 1\u20136","DOI":"10.1109\/ICME51207.2021.9428151"},{"key":"19368_CR31","doi-asserted-by":"publisher","unstructured":"Droste R, Jiao J, Noble JA (2020) Unified image and video saliency modeling. In: Vedaldi A, Bischof H, Brox T, et\u00a0al (eds) Computer Vision - ECCV 2020 - 16th European Conference, Glasgow, UK, August 23-28, 2020, Proceedings, Part V, Lecture Notes in Computer Science, vol 12350. Springer, pp 419\u2013435, https:\/\/doi.org\/10.1007\/978-3-030-58558-7_25","DOI":"10.1007\/978-3-030-58558-7_25"},{"issue":"10","key":"19368_CR32","doi-asserted-by":"crossref","first-page":"1380","DOI":"10.1038\/nn1537","volume":"8","author":"F Fang","year":"2005","unstructured":"Fang F, He S (2005) Cortical responses to invisible objects in the human dorsal and ventral pathways. Nat Neurosci 8(10):1380\u20131385","journal-title":"Nat Neurosci"},{"key":"19368_CR33","doi-asserted-by":"crossref","unstructured":"Fang J, Yan D, Qiao J et\u00a0al (2019) Dada-2000: Can driving accident be predicted by driver attentionf analyzed by a benchmark. In: 2019 IEEE Intelligent Transportation Systems Conference (ITSC), IEEE, pp 4303\u20134309","DOI":"10.1109\/ITSC.2019.8917218"},{"issue":"6","key":"19368_CR34","doi-asserted-by":"crossref","first-page":"4959","DOI":"10.1109\/TITS.2020.3044678","volume":"23","author":"J Fang","year":"2021","unstructured":"Fang J, Yan D, Qiao J et al (2021) Dada: Driver attention prediction in driving accident scenarios. IEEE Trans Intell Transp Syst 23(6):4959\u20134971","journal-title":"IEEE Trans Intell Transp Syst"},{"key":"19368_CR35","doi-asserted-by":"crossref","unstructured":"Fu R, Huang T, Li M et al (2023) A multimodal deep neural network for prediction of the driver\u2019s focus of attention based on anthropomorphic attention mechanism and prior knowledge. Expert Syst Appl 214:119157","DOI":"10.1016\/j.eswa.2022.119157"},{"key":"19368_CR36","doi-asserted-by":"crossref","unstructured":"Fu Z, Liu Q, Fu Z, et\u00a0al (2021) Stmtrack: Template-free visual tracking with space-time memory networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 13774\u201313783","DOI":"10.1109\/CVPR46437.2021.01356"},{"key":"19368_CR37","doi-asserted-by":"publisher","unstructured":"Gan S, Li Q, Wang Q, et\u00a0al (2021) Constructing personalized situation awareness dataset for hazard perception, comprehension, projection, and action of drivers. In: 24th IEEE International Intelligent Transportation Systems Conference, ITSC 2021, Indianapolis, IN, USA, September 19-22, 2021. IEEE, pp 1697\u20131704, https:\/\/doi.org\/10.1109\/ITSC48978.2021.9564543","DOI":"10.1109\/ITSC48978.2021.9564543"},{"issue":"11","key":"19368_CR38","doi-asserted-by":"crossref","first-page":"20912","DOI":"10.1109\/TITS.2022.3177640","volume":"23","author":"S Gan","year":"2022","unstructured":"Gan S, Pei X, Ge Y et al (2022) Multisource adaption for driver attention prediction in arbitrary driving scenes. IEEE Trans Intell Transp Syst 23(11):20912\u201320925","journal-title":"IEEE Trans Intell Transp Syst"},{"key":"19368_CR39","doi-asserted-by":"crossref","unstructured":"Giang WC, Shanti I, Chen HYW, et\u00a0al (2015) Smartwatches vs. smartphones: A preliminary report of driver behavior and perceived risk while responding to notifications. In: Proceedings of the 7th international conference on automotive user interfaces and interactive vehicular applications, pp 154\u2013161","DOI":"10.1145\/2799250.2799282"},{"key":"19368_CR40","doi-asserted-by":"crossref","unstructured":"Girma A, Amsalu S, Workineh A, et\u00a0al (2020) Deep learning with attention mechanism for predicting driver intention at intersection. In: 2020 IEEE Intelligent Vehicles Symposium (IV), IEEE, pp 1183\u20131188","DOI":"10.1109\/IV47402.2020.9304785"},{"issue":"6","key":"19368_CR41","doi-asserted-by":"crossref","first-page":"8268","DOI":"10.1007\/s11227-021-04151-2","volume":"78","author":"C Gou","year":"2022","unstructured":"Gou C, Zhou Y, Li D (2022) Driver attention prediction based on convolution and transformers. J Supercomput 78(6):8268\u20138284","journal-title":"J Supercomput"},{"issue":"4","key":"19368_CR42","doi-asserted-by":"crossref","first-page":"130","DOI":"10.1109\/MSP.2017.2699039","volume":"34","author":"JH Hansen","year":"2017","unstructured":"Hansen JH, Busso C, Zheng Y et al (2017) Driver modeling for detection and assessment of driver distraction: Examples from the utdrive test bed. IEEE Signal Process Mag 34(4):130\u2013142","journal-title":"IEEE Signal Process Mag"},{"key":"19368_CR43","doi-asserted-by":"crossref","unstructured":"Harel J, Koch C, Perona P (2006) Graph-based visual saliency. Adv Neural Inf Process Syst 19","DOI":"10.7551\/mitpress\/7503.003.0073"},{"key":"19368_CR44","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, et\u00a0al (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"19368_CR45","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, et\u00a0al (2016) Identity mappings in deep residual networks. In: Computer Vision\u2013ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, Part IV 14, Springer, pp 630\u2013645","DOI":"10.1007\/978-3-319-46493-0_38"},{"issue":"6","key":"19368_CR46","doi-asserted-by":"crossref","first-page":"813","DOI":"10.1111\/j.1467-7687.2009.00944.x","volume":"13","author":"S Hoehl","year":"2010","unstructured":"Hoehl S, Striano T (2010) The development of emotional face and eye gaze processing. Dev Sci 13(6):813\u2013825","journal-title":"Dev Sci"},{"issue":"1","key":"19368_CR47","first-page":"194","volume":"34","author":"X Hou","year":"2011","unstructured":"Hou X, Harel J, Koch C (2011) Image signature: Highlighting sparse salient regions. IEEE Trans Pattern Anal Mach Intell 34(1):194\u2013201","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"19368_CR48","doi-asserted-by":"crossref","first-page":"115697","DOI":"10.1016\/j.image.2019.115697","volume":"81","author":"Y Hu","year":"2020","unstructured":"Hu Y, Lu M, Lu X (2020) Feature refinement for image-based driver action recognition via multi-scale attention convolutional neural network. Signal Process Image Commun 81:115697","journal-title":"Signal Process Image Commun"},{"issue":"2","key":"19368_CR49","doi-asserted-by":"crossref","first-page":"1800","DOI":"10.1109\/TIE.2021.3057033","volume":"69","author":"Z Hu","year":"2021","unstructured":"Hu Z, Lv C, Hang P et al (2021) Data-driven estimation of driver attention using calibration-free eye gaze and scene features. IEEE Trans Ind Electron 69(2):1800\u20131808","journal-title":"IEEE Trans Ind Electron"},{"issue":"12","key":"19368_CR50","doi-asserted-by":"crossref","first-page":"24343","DOI":"10.1109\/TITS.2022.3208004","volume":"23","author":"Z Hu","year":"2022","unstructured":"Hu Z, Zhang Y, Li Q et al (2022) A novel heterogeneous network for modeling driver attention with multi-level visual content. IEEE Trans Intell Transp Syst 23(12):24343\u201324354","journal-title":"IEEE Trans Intell Transp Syst"},{"issue":"10","key":"19368_CR51","doi-asserted-by":"publisher","first-page":"19374","DOI":"10.1109\/TITS.2022.3166208","volume":"23","author":"T Huang","year":"2022","unstructured":"Huang T, Fu R (2022) Driver distraction detection based on the true driver\u2019s focus of attention. IEEE Trans Intell Transp Syst 23(10):19374\u201319386. https:\/\/doi.org\/10.1109\/TITS.2022.3166208","journal-title":"IEEE Trans Intell Transp Syst"},{"key":"19368_CR52","doi-asserted-by":"publisher","first-page":"109006","DOI":"10.1016\/j.knosys.2022.109006","volume":"251","author":"T Huang","year":"2022","unstructured":"Huang T, Fu R (2022) Prediction of the driver\u2019s focus of attention based on feature visualization of a deep autonomous driving model. Knowl Based Syst 251:109006. https:\/\/doi.org\/10.1016\/j.knosys.2022.109006","journal-title":"Knowl Based Syst"},{"issue":"10","key":"19368_CR53","doi-asserted-by":"crossref","first-page":"1179","DOI":"10.3390\/electronics8101179","volume":"8","author":"T Huang","year":"2019","unstructured":"Huang T, Zhao S, Geng L et al (2019) Unsupervised monocular depth estimation based on residual neural network of coarse-refined feature extractions for drone. Electronics 8(10):1179","journal-title":"Electronics"},{"key":"19368_CR54","doi-asserted-by":"crossref","unstructured":"Huang X, Shen C, Boix X, et\u00a0al (2015) Salicon: Reducing the semantic gap in saliency prediction by adapting deep neural networks. In: Proceedings of the IEEE international conference on computer vision, pp 262\u2013270","DOI":"10.1109\/ICCV.2015.38"},{"key":"19368_CR55","unstructured":"Ioffe S, Szegedy C (2015) Batch normalization: Accelerating deep network training by reducing internal covariate shift. In: Bach FR, Blei DM (eds) Proceedings of the 32nd International Conference on Machine Learning, ICML 2015, Lille, France, 6-11 July 2015, JMLR Workshop and Conference Proceedings, vol\u00a037. JMLR.org, pp 448\u2013456, http:\/\/proceedings.mlr.press\/v37\/ioffe15.html"},{"key":"19368_CR56","doi-asserted-by":"crossref","unstructured":"Isola P, Zhu JY, Zhou T, et\u00a0al (2017) Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1125\u20131134","DOI":"10.1109\/CVPR.2017.632"},{"issue":"6","key":"19368_CR57","doi-asserted-by":"crossref","first-page":"843","DOI":"10.1016\/j.neubiorev.2009.02.004","volume":"33","author":"RJ Itier","year":"2009","unstructured":"Itier RJ, Batty M (2009) Neural bases of eye and gaze processing: the core of social cognition. Neurosci Biobehav Rev 33(6):843\u2013863","journal-title":"Neurosci Biobehav Rev"},{"issue":"11","key":"19368_CR58","doi-asserted-by":"crossref","first-page":"1254","DOI":"10.1109\/34.730558","volume":"20","author":"L Itti","year":"1998","unstructured":"Itti L, Koch C, Niebur E (1998) A model of saliency-based visual attention for rapid scene analysis. IEEE Trans Pattern Anal Mach. Intell 20(11):1254\u20131259","journal-title":"IEEE Trans Pattern Anal Mach. Intell"},{"issue":"2","key":"19368_CR59","doi-asserted-by":"crossref","first-page":"1918","DOI":"10.1109\/JSEN.2020.3019258","volume":"21","author":"I Jegham","year":"2020","unstructured":"Jegham I, Khalifa AB, Alouani I et al (2020) Soft spatial attention-based multimodal driver action recognition using deep learning. IEEE Sens J 21(2):1918\u20131925","journal-title":"IEEE Sens J"},{"key":"19368_CR60","doi-asserted-by":"publisher","unstructured":"Jocher G, Chaurasia A, Stoken A, et\u00a0al (2022) ultralytics\/yolov5: v7.0 - YOLOv5 SOTA Realtime Instance Segmentation. https:\/\/doi.org\/10.5281\/zenodo.7347926","DOI":"10.5281\/zenodo.7347926"},{"key":"19368_CR61","unstructured":"Judd T, Durand F, Torralba A (2012) A benchmark of computational models of saliency to predict human fixations. MIT computer science and artificial intelligence laboratory technical report (Retrieved from https:\/\/www.hdlhandlenet\/17211\/68590)"},{"issue":"11","key":"19368_CR62","doi-asserted-by":"crossref","first-page":"5406","DOI":"10.3390\/app12115406","volume":"12","author":"K Kalda","year":"2022","unstructured":"Kalda K, Pizzagalli SL, Soe RM et al (2022) Language of driving for autonomous vehicles. Appl Sci 12(11):5406","journal-title":"Appl Sci"},{"issue":"7","key":"19368_CR63","doi-asserted-by":"crossref","first-page":"2030","DOI":"10.3390\/s20072030","volume":"20","author":"B Kang","year":"2020","unstructured":"Kang B, Lee Y (2020) High-resolution neural network for driver visual attention prediction. Sensors 20(7):2030","journal-title":"Sensors"},{"key":"19368_CR64","doi-asserted-by":"crossref","unstructured":"Kang B, Lee Y (2021) A driver\u2019s visual attention prediction using optical flow. Sensors 21(11):3722","DOI":"10.3390\/s21113722"},{"key":"19368_CR65","doi-asserted-by":"crossref","unstructured":"Kasahara I, Stent S, Park HS (2022) Look both ways: Self-supervising driver gaze estimation and road scene saliency. In: Computer Vision\u2013ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23\u201327, 2022, Proceedings, Part XIII, Springer, pp 126\u2013142","DOI":"10.1007\/978-3-031-19778-9_8"},{"key":"19368_CR66","unstructured":"Kay W, Carreira J, Simonyan K, et\u00a0al (2017) The kinetics human action video dataset. arXiv preprint arXiv:1705.06950"},{"key":"19368_CR67","doi-asserted-by":"publisher","unstructured":"Kellnhofer P, Recasens A, Stent S, et\u00a0al (2019) Gaze360: Physically unconstrained gaze estimation in the wild. In: 2019 IEEE\/CVF International Conference on Computer Vision, ICCV 2019, Seoul, Korea (South), October 27 - November 2, 2019. IEEE, pp 6911\u20136920, https:\/\/doi.org\/10.1109\/ICCV.2019.00701","DOI":"10.1109\/ICCV.2019.00701"},{"key":"19368_CR68","doi-asserted-by":"crossref","unstructured":"Kim J, Rohrbach A, Darrell T, et\u00a0al (2018) Textual explanations for self-driving vehicles. In: Proceedings of the European conference on computer vision (ECCV), pp 563\u2013578","DOI":"10.1007\/978-3-030-01216-8_35"},{"key":"19368_CR69","doi-asserted-by":"crossref","unstructured":"Kim J, Ma M, Kim K, et\u00a0al (2019) Progressive attention memory network for movie story question answering. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 8337\u20138346","DOI":"10.1109\/CVPR.2019.00853"},{"key":"19368_CR70","unstructured":"Kotseruba I, Tsotsos JK (2021) Behavioral research and practical models of drivers\u2019 attention. CoRR abs\/2104.05677. https:\/\/arxiv.org\/abs\/2104.05677,2104.05677"},{"issue":"6","key":"19368_CR71","doi-asserted-by":"crossref","first-page":"3420","DOI":"10.1109\/TITS.2020.2986697","volume":"22","author":"SM Kouchak","year":"2020","unstructured":"Kouchak SM, Gaffar A (2020) Detecting driver behavior using stacked long short term memory network with attention layer. IEEE Trans Intell Transp Syst 22(6):3420\u20133429","journal-title":"IEEE Trans Intell Transp Syst"},{"issue":"6","key":"19368_CR72","doi-asserted-by":"crossref","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky A, Sutskever I, Hinton GE (2017) Imagenet classification with deep convolutional neural networks. Communications of the ACM 60(6):84\u201390","journal-title":"Communications of the ACM"},{"key":"19368_CR73","doi-asserted-by":"publisher","first-page":"261","DOI":"10.1016\/j.neunet.2020.05.004","volume":"129","author":"A Kroner","year":"2020","unstructured":"Kroner A, Senden M, Driessens K et al (2020) Contextual encoder-decoder network for visual saliency prediction. Neural Netw 129:261\u2013270. https:\/\/doi.org\/10.1016\/j.neunet.2020.05.004","journal-title":"Neural Netw"},{"key":"19368_CR74","unstructured":"K\u00fcmmerer M, Wallis T, Bethge M (2014) How close are we to understanding image-based saliency? arXiv preprint arXiv:1409.7686"},{"issue":"52","key":"19368_CR75","doi-asserted-by":"crossref","first-page":"16054","DOI":"10.1073\/pnas.1510393112","volume":"112","author":"M K\u00fcmmerer","year":"2015","unstructured":"K\u00fcmmerer M, Wallis TS, Bethge M (2015) Information-theoretic model comparison unifies saliency metrics. Proc Natl Acad Sci 112(52):16054\u201316059","journal-title":"Proc Natl Acad Sci"},{"key":"19368_CR76","doi-asserted-by":"crossref","unstructured":"Kwon YH, Park MG (2019) Predicting future frames using retrospective cycle gan. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 1811\u20131820","DOI":"10.1109\/CVPR.2019.00191"},{"issue":"19","key":"19368_CR77","doi-asserted-by":"crossref","first-page":"2483","DOI":"10.1016\/j.visres.2007.06.015","volume":"47","author":"O Le Meur","year":"2007","unstructured":"Le Meur O, Le Callet P, Barba D (2007) Predicting visual fixations on video based on low-level visual features. Vis Res 47(19):2483\u20132498","journal-title":"Vis Res"},{"key":"19368_CR78","doi-asserted-by":"crossref","unstructured":"Li C, Wand M (2016) Precomputed real-time texture synthesis with markovian generative adversarial networks. In: European conference on computer vision, Springer, pp 702\u2013716","DOI":"10.1007\/978-3-319-46487-9_43"},{"issue":"4","key":"19368_CR79","doi-asserted-by":"crossref","first-page":"996","DOI":"10.1109\/TPAMI.2012.147","volume":"35","author":"J Li","year":"2012","unstructured":"Li J, Levine MD, An X et al (2012) Visual saliency based on scale-space analysis in the frequency domain. IEEE Trans Pattern Anal Mach Intell 35(4):996\u20131010","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"19368_CR80","doi-asserted-by":"crossref","unstructured":"Li Q, Liu C, Chang F, et\u00a0al (2022) Adaptive short-temporal induced aware fusion network for predicting attention regions like a driver. IEEE Trans Intell Transp Syst","DOI":"10.1109\/TITS.2022.3165619"},{"key":"19368_CR81","doi-asserted-by":"crossref","first-page":"83138","DOI":"10.1109\/ACCESS.2022.3197146","volume":"10","author":"T Li","year":"2022","unstructured":"Li T, Zhang Y, Li Q et al (2022) Ab-dlm: An improved deep learning model based on attention mechanism and bifpn for driver distraction behavior detection. IEEE Access 10:83138\u201383151","journal-title":"IEEE Access"},{"issue":"2","key":"19368_CR82","doi-asserted-by":"crossref","first-page":"197","DOI":"10.1109\/MITS.2021.3049404","volume":"14","author":"L Lin","year":"2021","unstructured":"Lin L, Li W, Bi H et al (2021) Vehicle trajectory prediction using lstms with spatial-temporal attention mechanisms. IEEE Intell Transp Syst Mag 14(2):197\u2013208","journal-title":"IEEE Intell Transp Syst Mag"},{"issue":"9","key":"19368_CR83","doi-asserted-by":"crossref","first-page":"4191","DOI":"10.3390\/app12094191","volume":"12","author":"Y Lin","year":"2022","unstructured":"Lin Y, Cao D, Fu Z et al (2022) A lightweight attention-based network towards distracted driving behavior recognition. Appl Sci 12(9):4191","journal-title":"Appl Sci"},{"key":"19368_CR84","unstructured":"Linardos P, Mohedano E, Nieto JJ, et\u00a0al (2019) Simple vs complex temporal recurrences for video saliency prediction. arXiv preprint arXiv:1907.01869"},{"key":"19368_CR85","doi-asserted-by":"crossref","unstructured":"Lisheng J, Bingdong J, Baicang G et\u00a0al (2022) Mtsf: Multi-scale temporal-spatial fusion network for driver attention prediction. Available at SSRN: https:\/\/www.ssrncom\/abstract=4167535","DOI":"10.2139\/ssrn.4167535"},{"key":"19368_CR86","doi-asserted-by":"crossref","unstructured":"Liu Z, Mao H, Wu CY, et\u00a0al (2022) A convnet for the 2020s. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 11976\u201311986","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"19368_CR87","unstructured":"Lou J, Lin H, Marshall D, et\u00a0al (2021) Transalnet: Visual saliency prediction using transformers. CoRR abs\/2110.03593. https:\/\/arxiv.org\/abs\/2110.03593, 2110.03593"},{"key":"19368_CR88","doi-asserted-by":"crossref","first-page":"4198","DOI":"10.1109\/TMM.2020.3038311","volume":"23","author":"K Lv","year":"2020","unstructured":"Lv K, Sheng H, Xiong Z et al (2020) Improving driver gaze prediction with reinforced attention. IEEE Trans Multimedia 23:4198\u20134207","journal-title":"IEEE Trans Multimedia"},{"issue":"3","key":"19368_CR89","doi-asserted-by":"crossref","first-page":"1145","DOI":"10.3390\/app12031145","volume":"12","author":"E Mag\u00e1n","year":"2022","unstructured":"Mag\u00e1n E, Sesmero MP, Alonso-Weber JM et al (2022) Driver drowsiness detection by applying deep learning techniques to sequences of images. Appl Sci 12(3):1145","journal-title":"Appl Sci"},{"key":"19368_CR90","doi-asserted-by":"crossref","unstructured":"Min K, Corso JJ (2019) Tased-net: Temporally-aggregating spatial encoder-decoder network for video saliency detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 2394\u20132403","DOI":"10.1109\/ICCV.2019.00248"},{"key":"19368_CR91","doi-asserted-by":"crossref","first-page":"3805","DOI":"10.1109\/TIP.2020.2966082","volume":"29","author":"X Min","year":"2020","unstructured":"Min X, Zhai G, Zhou J et al (2020) A multimodal saliency model for videos with high audio-visual correspondence. IEEE Trans Image Process 29:3805\u20133819","journal-title":"IEEE Trans Image Process"},{"key":"19368_CR92","unstructured":"Oktay O, Schlemper J, Folgoc LL, et\u00a0al (2018) Attention u-net: Learning where to look for the pancreas. CoRR abs\/1804.03999. http:\/\/arxiv.org\/abs\/1804.03999, 1804.03999"},{"key":"19368_CR93","doi-asserted-by":"crossref","first-page":"239","DOI":"10.1016\/j.aap.2018.04.018","volume":"117","author":"C Ortiz","year":"2018","unstructured":"Ortiz C, Ortiz-Peregrina S, Castro J et al (2018) Driver distraction by smartphone use (whatsapp) in different age groups. Accid Anal Prev 117:239\u2013249","journal-title":"Accid Anal Prev"},{"issue":"7","key":"19368_CR94","doi-asserted-by":"crossref","first-page":"1720","DOI":"10.1109\/TPAMI.2018.2845370","volume":"41","author":"A Palazzi","year":"2018","unstructured":"Palazzi A, Abati D, Solera F et al (2018) Predicting the driver\u2019s focus of attention: the dr (eye) ve project. IEEE Trans Pattern Anal Mach Intell 41(7):1720\u20131733","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"19368_CR95","unstructured":"Pan J, Ferrer CC, McGuinness K, et\u00a0al (2017) Salgan: Visual saliency prediction with generative adversarial networks. arXiv preprint arXiv:1701.01081"},{"key":"19368_CR96","doi-asserted-by":"publisher","unstructured":"Pan X, Ge C, Lu R, et\u00a0al (2022) On the integration of self-attention and convolution. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022. IEEE, pp 805\u2013815, https:\/\/doi.org\/10.1109\/CVPR52688.2022.00089","DOI":"10.1109\/CVPR52688.2022.00089"},{"key":"19368_CR97","doi-asserted-by":"crossref","unstructured":"Pandey NN, Muppalaneni NB (2022) A survey on visual and non-visual features in driver\u2019s drowsiness detection. Multimed Tools Appl 81(26):38175\u201338215","DOI":"10.1007\/s11042-022-13150-1"},{"key":"19368_CR98","doi-asserted-by":"crossref","unstructured":"Pang Y, Zhao X, Zhang L, et\u00a0al (2020) Multi-scale interactive network for salient object detection. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 9413\u20139422","DOI":"10.1109\/CVPR42600.2020.00943"},{"key":"19368_CR99","doi-asserted-by":"publisher","unstructured":"Pele O, Werman M (2008) A linear time histogram metric for improved SIFT matching. In: Forsyth DA, Torr PHS, Zisserman A (eds) Computer Vision - ECCV 2008, 10th European Conference on Computer Vision, Marseille, France, October 12-18, 2008, Proceedings, Part III, Lecture Notes in Computer Science, vol 5304. Springer, pp 495\u2013508, https:\/\/doi.org\/10.1007\/978-3-540-88690-7_37","DOI":"10.1007\/978-3-540-88690-7_37"},{"key":"19368_CR100","doi-asserted-by":"crossref","unstructured":"Peng X, Zhao A, Wang S, et\u00a0al (2019) Attention-driven driving maneuver detection system. In: 2019 International Joint Conference on Neural Networks (IJCNN), IEEE, pp 1\u20138","DOI":"10.1109\/IJCNN.2019.8852404"},{"key":"19368_CR101","doi-asserted-by":"crossref","first-page":"8","DOI":"10.1016\/j.apergo.2018.09.001","volume":"75","author":"D Perlman","year":"2019","unstructured":"Perlman D, Samost A, Domel AG et al (2019) The relative impact of smartwatch and smartphone use while driving on workload, attention, and driving performance. Appl Ergon 75:8\u201316","journal-title":"Appl Ergon"},{"issue":"18","key":"19368_CR102","doi-asserted-by":"crossref","first-page":"2397","DOI":"10.1016\/j.visres.2005.03.019","volume":"45","author":"RJ Peters","year":"2005","unstructured":"Peters RJ, Iyer A, Itti L et al (2005) Components of bottom-up gaze allocation in natural images. Vis Res 45(18):2397\u20132416","journal-title":"Vis Res"},{"issue":"1","key":"19368_CR103","doi-asserted-by":"crossref","first-page":"3","DOI":"10.1080\/00335558008248231","volume":"32","author":"MI Posner","year":"1980","unstructured":"Posner MI (1980) Orienting of attention. Q J Exp Psychol 32(1):3\u201325","journal-title":"Q J Exp Psychol"},{"issue":"4","key":"19368_CR104","doi-asserted-by":"crossref","first-page":"840","DOI":"10.1518\/001872005775570961","volume":"47","author":"AK Pradhan","year":"2005","unstructured":"Pradhan AK, Hammel KR, DeRamus R et al (2005) Using eye movements to evaluate effects of driver age on risk perception in a driving simulator. Hum Factors 47(4):840\u2013852","journal-title":"Hum Factors"},{"key":"19368_CR105","doi-asserted-by":"crossref","first-page":"107404","DOI":"10.1016\/j.patcog.2020.107404","volume":"106","author":"X Qin","year":"2020","unstructured":"Qin X, Zhang Z, Huang C et al (2020) U2-net: Going deeper with nested u-structure for salient object detection. Pattern Recognit 106:107404","journal-title":"Pattern Recognit"},{"key":"19368_CR106","unstructured":"Ramos J (2022) Autonomous vehicles and accidents: are they safer than vehicles operated by drivers? https:\/\/tomorrow.city\/a\/self-driving-car-accident-rate"},{"key":"19368_CR107","doi-asserted-by":"publisher","unstructured":"Reddy N, Jain S, Yarlagadda P, et\u00a0al (2020) Tidying deep saliency prediction architectures. In: IEEE\/RSJ International Conference on Intelligent Robots and Systems, IROS 2020, Las Vegas, NV, USA, October 24, 2020 - January 24, 2021. IEEE, pp 10241\u201310247, https:\/\/doi.org\/10.1109\/IROS45743.2020.9341574","DOI":"10.1109\/IROS45743.2020.9341574"},{"key":"19368_CR108","doi-asserted-by":"crossref","unstructured":"Rong Y, Kassautzki NR, Fuhl W, et\u00a0al (2022) Where and what: Driver attention-based object detection. Proc ACM Hum Comput Interact 6(ETRA):1\u201322","DOI":"10.1145\/3530887"},{"key":"19368_CR109","doi-asserted-by":"crossref","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. In: International Conference on Medical image computing and computer-assisted intervention, Springer, pp 234\u2013241","DOI":"10.1007\/978-3-319-24574-4_28"},{"issue":"2","key":"19368_CR110","doi-asserted-by":"crossref","first-page":"99","DOI":"10.1023\/A:1026543900054","volume":"40","author":"Y Rubner","year":"2000","unstructured":"Rubner Y, Tomasi C, Guibas LJ (2000) The earth mover\u2019s distance as a metric for image retrieval. Int J Comput Vis 40(2):99","journal-title":"Int J Comput Vis"},{"key":"19368_CR111","doi-asserted-by":"crossref","unstructured":"Rundo F, Spampinato C, Battiato S, et\u00a0al (2020) Advanced 1d temporal deep dilated convolutional embedded perceptual system for fast car-driver drowsiness monitoring. In: 2020 AEIT International Conference of Electrical and Electronic Technologies for Automotive (AEIT AUTOMOTIVE), IEEE, pp 1\u20136","DOI":"10.23919\/AEITAUTOMOTIVE50086.2020.9307400"},{"key":"19368_CR112","unstructured":"Shao H, Wang L, Chen R, et\u00a0al (2022) Safety-enhanced autonomous driving using interpretable sensor fusion transformer. arXiv preprint arXiv:2207.14024"},{"key":"19368_CR113","doi-asserted-by":"crossref","unstructured":"Sharma S, Kumar V (2023) Distracted driver detection using learning representations. Multimed Tools Appl pp 1\u201318","DOI":"10.1007\/s11042-023-14635-3"},{"key":"19368_CR114","doi-asserted-by":"crossref","unstructured":"Shi B, Dong W, Zhan Z (2022) Adafi-fcn: an adaptive feature integration fully convolutional network for predicting driver\u2019s visual attention. Geo-spatial Information Science pp 1\u201317","DOI":"10.1080\/10095020.2022.2147028"},{"key":"19368_CR115","doi-asserted-by":"crossref","first-page":"12","DOI":"10.1016\/j.trf.2015.03.009","volume":"31","author":"J Son","year":"2015","unstructured":"Son J, Park M, Park BB (2015) The effect of age, gender and roadway environment on the acceptance and effectiveness of advanced driver assistance systems. Transportation research part F: traffic psychology and behaviour 31:12\u201324","journal-title":"Transportation research part F: traffic psychology and behaviour"},{"issue":"1","key":"19368_CR116","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1007\/BF00130487","volume":"7","author":"MJ Swain","year":"1991","unstructured":"Swain MJ, Ballard DH (1991) Color indexing. Int J Comput Vis 7(1):11\u201332. https:\/\/doi.org\/10.1007\/BF00130487","journal-title":"Color indexing. Int J Comput Vis"},{"key":"19368_CR117","doi-asserted-by":"crossref","unstructured":"Takeda Y, Sato T, Kimura K et al (2016) Electrophysiological evaluation of attention in drivers and passengers: Toward an understanding of drivers\u2019 attentional state in autonomous vehicles. Transportation research part F: traffic psychology and behaviour 42:140\u2013150","DOI":"10.1016\/j.trf.2016.07.008"},{"key":"19368_CR118","unstructured":"Tavakoli HR, Borji A, Rahtu E, et\u00a0al (2019) Dave: A deep audio-visual embedding for dynamic saliency prediction. arXiv preprint arXiv:1905.10693"},{"key":"19368_CR119","doi-asserted-by":"crossref","unstructured":"Tawari A, Kang B (2017) A computational framework for driver\u2019s visual attention using a fully convolutional architecture. In: 2017 IEEE Intelligent Vehicles Symposium (IV), IEEE, pp 887\u2013894","DOI":"10.1109\/IVS.2017.7995828"},{"issue":"7","key":"19368_CR120","doi-asserted-by":"crossref","first-page":"1335","DOI":"10.1109\/JAS.2022.105716","volume":"9","author":"H Tian","year":"2022","unstructured":"Tian H, Deng T, Yan H (2022) Driving as well as on a sunny day? predicting driver\u2019s fixation in rainy weather conditions via a dual-branch visual model. IEEE\/CAA J Autom Sin 9(7):1335\u20131338","journal-title":"IEEE\/CAA J Autom Sin"},{"key":"19368_CR121","doi-asserted-by":"crossref","unstructured":"Tran D, Bourdev L, Fergus R, et\u00a0al (2015) Learning spatiotemporal features with 3d convolutional networks. In: Proceedings of the IEEE international conference on computer vision, pp 4489\u20134497","DOI":"10.1109\/ICCV.2015.510"},{"issue":"7","key":"19368_CR122","doi-asserted-by":"crossref","first-page":"4","DOI":"10.1167\/9.7.4","volume":"9","author":"PH Tseng","year":"2009","unstructured":"Tseng PH, Carmi R, Cameron IG et al (2009) Quantifying center bias of observers in free viewing of dynamic natural scenes. J Vis 9(7):4\u20134","journal-title":"J Vis"},{"key":"19368_CR123","doi-asserted-by":"crossref","unstructured":"Ulrich L, Nonis F, Vezzetti E et al (2021) Can adas distract driver\u2019s attention? an rgb-d camera and deep learning-based analysis. Appl Sci 11(24):11587","DOI":"10.3390\/app112411587"},{"key":"19368_CR124","doi-asserted-by":"crossref","unstructured":"Kastner S, Ungerleider LG (2000) Mechanisms of visual attention in the human cortex. Annu Rev Neurosci 23(1):315\u2013341","DOI":"10.1146\/annurev.neuro.23.1.315"},{"key":"19368_CR125","unstructured":"Vaswani A, Shazeer N, Parmar N, et\u00a0al (2017) Attention is all you need. Adv Neural Inf Process Syst 30"},{"key":"19368_CR126","doi-asserted-by":"crossref","unstructured":"Wang W, Shen J, Porikli F (2015) Saliency-aware geodesic video object segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3395\u20133402","DOI":"10.1109\/CVPR.2015.7298961"},{"issue":"1","key":"19368_CR127","doi-asserted-by":"crossref","first-page":"220","DOI":"10.1109\/TPAMI.2019.2924417","volume":"43","author":"W Wang","year":"2019","unstructured":"Wang W, Shen J, Xie J et al (2019) Revisiting video saliency prediction in the deep learning era. IEEE Trans Pattern Anal Mach Intell 43(1):220\u2013237","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"19368_CR128","doi-asserted-by":"crossref","unstructured":"Wang X, Gupta A (2018) Videos as space-time region graphs. In: Proceedings of the European conference on computer vision (ECCV), pp 399\u2013417","DOI":"10.1007\/978-3-030-01228-1_25"},{"issue":"50","key":"19368_CR129","doi-asserted-by":"crossref","first-page":"11666","DOI":"10.1523\/JNEUROSCI.3414-05.2005","volume":"25","author":"BS Webb","year":"2005","unstructured":"Webb BS, Dhruv NT, Solomon SG et al (2005) Early and late mechanisms of surround suppression in striate cortex of macaque. J Neurosci 25(50):11666\u201311675","journal-title":"J Neurosci"},{"key":"19368_CR130","doi-asserted-by":"crossref","unstructured":"Woo S, Park J, Lee JY, et\u00a0al (2018) Cbam: Convolutional block attention module. In: Proceedings of the European conference on computer vision (ECCV), pp 3\u201319","DOI":"10.1007\/978-3-030-01234-2_1"},{"issue":"1","key":"19368_CR131","first-page":"121","volume":"29","author":"GF Woodman","year":"2003","unstructured":"Woodman GF, Luck SJ (2003) Serial deployment of attention during visual search. J Exp Psychol: Hum Percept. Perform 29(1):121","journal-title":"J Exp Psychol: Hum Percept. Perform"},{"key":"19368_CR132","doi-asserted-by":"crossref","unstructured":"Wu CY, Krahenbuhl P (2021) Towards long-form video understanding. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 1884\u20131894","DOI":"10.1109\/CVPR46437.2021.00192"},{"key":"19368_CR133","doi-asserted-by":"crossref","unstructured":"Wu CY, Li Y, Mangalam K, et\u00a0al (2022) Memvit: Memory-augmented multiscale vision transformer for efficient long-term video recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 13587\u201313597","DOI":"10.1109\/CVPR52688.2022.01322"},{"issue":"7","key":"19368_CR134","doi-asserted-by":"crossref","first-page":"696","DOI":"10.1080\/15389588.2019.1644627","volume":"20","author":"L Wundersitz","year":"2019","unstructured":"Wundersitz L (2019) Driver distraction and inattention in fatal and injury crashes: Findings from in-depth road crash data. Traffic Inj Prev 20(7):696\u2013701","journal-title":"Traffic Inj Prev"},{"key":"19368_CR135","doi-asserted-by":"crossref","unstructured":"Xia Y, Zhang D, Kim J, et\u00a0al (2018) Predicting driver attention in critical situations. In: Asian conference on computer vision, Springer, pp 658\u2013674","DOI":"10.1007\/978-3-030-20873-8_42"},{"key":"19368_CR136","doi-asserted-by":"publisher","unstructured":"Xie C, Xia C, Ma M, et\u00a0al (2022) Pyramid grafting network for one-stage high resolution saliency detection. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022. IEEE, pp 11707\u201311716, https:\/\/doi.org\/10.1109\/CVPR52688.2022.01142","DOI":"10.1109\/CVPR52688.2022.01142"},{"key":"19368_CR137","doi-asserted-by":"publisher","unstructured":"Xu F, Li J, Yuan Q, et\u00a0al (2020) A dual-attention-based neural network for see-through driving decision. In: 92nd IEEE Vehicular Technology Conference, VTC Fall 2020, Victoria, BC, Canada, November 18 - December 16, 2020. IEEE, pp 1\u20136, https:\/\/doi.org\/10.1109\/VTC2020-Fall49728.2020.9348588","DOI":"10.1109\/VTC2020-Fall49728.2020.9348588"},{"key":"19368_CR138","doi-asserted-by":"publisher","unstructured":"Xu H, Gao Y, Yu F, et\u00a0al (2017) End-to-end learning of driving models from large-scale video datasets. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017. IEEE Computer Society, pp 3530\u20133538, https:\/\/doi.org\/10.1109\/CVPR.2017.376","DOI":"10.1109\/CVPR.2017.376"},{"issue":"1","key":"19368_CR139","doi-asserted-by":"crossref","first-page":"309","DOI":"10.3390\/app12010309","volume":"12","author":"F Yan","year":"2022","unstructured":"Yan F, Chen C, Xiao P et al (2022) Review of visual saliency prediction: Development process from neurobiological basis to deep models. Applied Sciences 12(1):309","journal-title":"Applied Sciences"},{"issue":"2","key":"19368_CR140","doi-asserted-by":"crossref","first-page":"221","DOI":"10.1109\/TIV.2022.3162719","volume":"7","author":"D Yang","year":"2022","unstructured":"Yang D, Zhang H, Yurtsever E et al (2022) Predicting pedestrian crossing intention with feature fusion and spatio-temporal attention. IEEE Trans Intell Transp Syst 7(2):221\u2013230","journal-title":"IEEE Trans Intell Transp Syst"},{"issue":"8","key":"19368_CR141","doi-asserted-by":"crossref","first-page":"2503","DOI":"10.1109\/TCSVT.2019.2925844","volume":"30","author":"P Yi","year":"2019","unstructured":"Yi P, Wang Z, Jiang K et al (2019) Multi-temporal ultra dense memory network for video super-resolution. IEEE Trans Circ Syst Video Tech 30(8):2503\u20132516","journal-title":"IEEE Trans Circ Syst Video Tech"},{"key":"19368_CR142","doi-asserted-by":"publisher","unstructured":"Yu C, Wang J, Peng C, et\u00a0al (2018) Bisenet: Bilateral segmentation network for real-time semantic segmentation. In: Ferrari V, Hebert M, Sminchisescu C, et\u00a0al (eds) Computer Vision - ECCV 2018 - 15th European Conference, Munich, Germany, September 8-14, 2018, Proceedings, Part XIII, Lecture Notes in Computer Science, vol 11217. Springer, pp 334\u2013349, https:\/\/doi.org\/10.1007\/978-3-030-01261-8_20","DOI":"10.1007\/978-3-030-01261-8_20"},{"issue":"1","key":"19368_CR143","doi-asserted-by":"crossref","first-page":"40","DOI":"10.3141\/2663-06","volume":"2663","author":"M Zahabi","year":"2017","unstructured":"Zahabi M, Pankok C Jr, Kaber DB et al (2017) On-road visual sign salience, driver attention allocation, and target detection accuracy. Transp Res Rec 2663(1):40\u201347","journal-title":"Transp Res Rec"},{"key":"19368_CR144","doi-asserted-by":"crossref","unstructured":"Zatsarynna O, Abu\u00a0Farha Y, Gall J (2021) Multi-modal temporal convolutional network for anticipating actions in egocentric videos. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 2249\u20132258","DOI":"10.1109\/CVPRW53098.2021.00254"},{"key":"19368_CR145","unstructured":"Zeng Y, Zhuge Y, Lu H, et\u00a0al (2019) Joint learning of saliency detection and weakly supervised semantic segmentation. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 7223\u20137233"},{"key":"19368_CR146","doi-asserted-by":"crossref","first-page":"1138","DOI":"10.1109\/TNSRE.2021.3089594","volume":"29","author":"G Zhang","year":"2021","unstructured":"Zhang G, Etemad A (2021) Capsule attention for multimodal eeg-eog representation learning with application to driver vigilance estimation. IEEE Trans Neural Syst Rehabil Eng 29:1138\u20131149","journal-title":"IEEE Trans Neural Syst Rehabil Eng"},{"key":"19368_CR147","unstructured":"Zhang H, Goodfellow I, Metaxas D, et\u00a0al (2019) Self-attention generative adversarial networks. In: International conference on machine learning, PMLR, pp 7354\u20137363"},{"key":"19368_CR148","doi-asserted-by":"crossref","unstructured":"Zhang X, Park S, Beeler T, et\u00a0al (2020) Eth-xgaze: A large scale dataset for gaze estimation under extreme head pose and gaze variation. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part V 16, Springer, pp 365\u2013381","DOI":"10.1007\/978-3-030-58558-7_22"},{"key":"19368_CR149","doi-asserted-by":"crossref","first-page":"101280","DOI":"10.1016\/j.aei.2021.101280","volume":"48","author":"L Zhao","year":"2021","unstructured":"Zhao L, Yang F, Bu L et al (2021) Driver behavior detection via adaptive spatial attention mechanism. Adv Eng Inform 48:101280","journal-title":"Adv Eng Inform"},{"key":"19368_CR150","doi-asserted-by":"crossref","unstructured":"Zhao S, Han G, Zhao Q et al (2020) Prediction of driver\u2019s attention points based on attention model. Appl Sci 10(3):1083","DOI":"10.3390\/app10031083"},{"key":"19368_CR151","doi-asserted-by":"publisher","unstructured":"Zhao T, Wu X (2019) Pyramid feature attention network for saliency detection. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16-20, 2019. Computer Vision Foundation \/ IEEE, pp 3085\u20133094, https:\/\/doi.org\/10.1109\/CVPR.2019.00320, http:\/\/openaccess.thecvf.com\/content_CVPR_2019\/html\/Zhao_Pyramid_Feature_Attention_Network_for_Saliency_Detection_CVPR_2019_paper.html","DOI":"10.1109\/CVPR.2019.00320"},{"key":"19368_CR152","doi-asserted-by":"crossref","unstructured":"Zhou X, Koltun V, Kr\u00e4henb\u00fchl P (2020) Tracking objects as points. In: European Conference on Computer Vision, Springer, pp 474\u2013490","DOI":"10.1007\/978-3-030-58548-8_28"},{"key":"19368_CR153","doi-asserted-by":"crossref","unstructured":"Zhu D, Zhao D, Min X, et\u00a0al (2021) Lavs: A lightweight audio-visual saliency prediction model. In: 2021 IEEE International Conference on Multimedia and Expo (ICME), IEEE, pp 1\u20136","DOI":"10.1109\/ICME51207.2021.9428415"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-19368-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-024-19368-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-19368-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,1]],"date-time":"2025-05-01T05:03:49Z","timestamp":1746075829000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-024-19368-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,18]]},"references-count":153,"journal-issue":{"issue":"12","published-online":{"date-parts":[[2025,4]]}},"alternative-id":["19368"],"URL":"https:\/\/doi.org\/10.1007\/s11042-024-19368-5","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2024,5,18]]},"assertion":[{"value":"10 August 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 March 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 May 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 May 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that there are no conflicts of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of Interest"}}]}}