{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T13:39:46Z","timestamp":1740145186366,"version":"3.37.3"},"reference-count":36,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2024,9,11]],"date-time":"2024-09-11T00:00:00Z","timestamp":1726012800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,9,11]],"date-time":"2024-09-11T00:00:00Z","timestamp":1726012800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62001173","62001173","62001173","62233013","62001173"],"award-info":[{"award-number":["62001173","62001173","62001173","62233013","62001173"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Real-Time Image Proc"],"published-print":{"date-parts":[[2024,10]]},"DOI":"10.1007\/s11554-024-01549-y","type":"journal-article","created":{"date-parts":[[2024,9,11]],"date-time":"2024-09-11T11:09:18Z","timestamp":1726052958000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Mfdd: Multi-scale attention fatigue and distracted driving detector based on facial features"],"prefix":"10.1007","volume":"21","author":[{"given":"Yulin","family":"Shi","sequence":"first","affiliation":[]},{"given":"Jintao","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Xingming","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Jiehao","family":"Luo","sequence":"additional","affiliation":[]},{"given":"Xiaoyu","family":"Tang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,9,11]]},"reference":[{"unstructured":"Prat,C.S., Seo,R., Yamasaki,B.L.: The role of individual differences in working memory capacity on reading comprehension ability,in Handbook of Individual Differences in Reading (Routledge, 2015), 331\u2013347","key":"1549_CR1"},{"key":"1549_CR2","doi-asserted-by":"publisher","DOI":"10.1016\/j.trc.2020.102615","volume":"115","author":"Y Xing","year":"2020","unstructured":"Xing, Y., Lv, C., Wang, H., Cao, D., Velenis, E.: An ensemble deep learning approach for driver lane change intention inference. Transp. Res. Part C: Emerg Technol. 115, 102615 (2020)","journal-title":"Transp. Res. Part C: Emerg Technol."},{"doi-asserted-by":"crossref","unstructured":"Kapoor,K., Pamula,R., Murthy,S.V.: Real-time driver distraction detection system using convolutional neural networks, In: Proceedings of ICETIT 2019: Emerging Trends in Information Technology, (2020), 280\u2013291","key":"1549_CR3","DOI":"10.1007\/978-3-030-30577-2_24"},{"issue":"11","key":"1549_CR4","doi-asserted-by":"publisher","first-page":"2278","DOI":"10.1109\/5.726791","volume":"86","author":"Y LeCun","year":"1998","unstructured":"LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proceed. IEEE 86(11), 2278 (1998)","journal-title":"Proceed. IEEE"},{"issue":"6","key":"1549_CR5","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, E.: ImageNet classification with deep convolutional neural networks. Commun. ACM 60(6), 84 (2017)","journal-title":"Commun. ACM"},{"doi-asserted-by":"crossref","unstructured":"He,K., Zhang,X., Ren,S., Sun,J.: Deep residual learning for image recognition, In: Proceedings of the IEEE conference on computer vision and pattern recognition (2016), 770\u2013778","key":"1549_CR6","DOI":"10.1109\/CVPR.2016.90"},{"key":"1549_CR7","first-page":"1140","volume":"35","author":"MH Guo","year":"2022","unstructured":"Guo, M.H., Lu, C.Z., Hou, Q., Liu, Z., Cheng, M.M., Hu, S.M.: Segnext: Rethinking convolutional attention design for semantic segmentation. Adv. Neural. Inf. Process. Syst. 35, 1140 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"doi-asserted-by":"crossref","unstructured":"Ding,X., Zhang,X., Han,J., Ding,G.: Scaling up your kernels to 31x31: Revisiting large kernel design in cnns, In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (2022), 11,963\u201311,975","key":"1549_CR8","DOI":"10.1109\/CVPR52688.2022.01166"},{"doi-asserted-by":"crossref","unstructured":"Sandler,M., Howard,A., Zhu,M., Zhmoginov,A., Chen,L.C.: Mobilenetv2: Inverted residuals and linear bottlenecks, In: Proceedings of the IEEE conference on computer vision and pattern recognition (2018), 4510\u20134520","key":"1549_CR9","DOI":"10.1109\/CVPR.2018.00474"},{"doi-asserted-by":"crossref","unstructured":"Hu,J., Shen,L., Sun,G.: Squeeze-and-excitation networks, In: Proceedings of the IEEE conference on computer vision and pattern recognition (2018), 7132\u20137141","key":"1549_CR10","DOI":"10.1109\/CVPR.2018.00745"},{"doi-asserted-by":"crossref","unstructured":"Li,Y., Li,X., Yang,J.: Spatial group-wise enhance: Enhancing semantic feature learning in cnn,in Proceedings of the Asian Conference on Computer Vision (2022), 687\u2013702","key":"1549_CR11","DOI":"10.1007\/978-3-031-26348-4_19"},{"doi-asserted-by":"crossref","unstructured":"Woo,S., Park,J., Lee,J.Y., Kweon,I.S.: Cbam: Convolutional block attention module, In: Proceedings of the European conference on computer vision (ECCV) (2018), 3\u201319","key":"1549_CR12","DOI":"10.1007\/978-3-030-01234-2_1"},{"doi-asserted-by":"crossref","unstructured":"Hou,Q., Zhou,D., Feng,J.: Coordinate attention for efficient mobile network design, In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (2021), 13,713\u201313,722","key":"1549_CR13","DOI":"10.1109\/CVPR46437.2021.01350"},{"unstructured":"Lin,M., Chen,Q., Yan,S.: Network in network, arXiv preprint arXiv:1312.4400 (2013)","key":"1549_CR14"},{"doi-asserted-by":"crossref","unstructured":"Lin,W., Wu,Z., Chen,J., Huang,J., Jin,L.: Scale-aware modulation meet transformer, In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (2023), 6015\u20136026","key":"1549_CR15","DOI":"10.1109\/ICCV51070.2023.00553"},{"key":"1549_CR16","first-page":"12077","volume":"34","author":"E Xie","year":"2021","unstructured":"Xie, E., Wang, W., Yu, Z., Anandkumar, A., Alvarez, J.M., Luo, P.: SegFormer: Simple and efficient design for semantic segmentation with transformers. Adv. Neural. Inf. Process. Syst. 34, 12077 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"4","key":"1549_CR17","first-page":"733","volume":"9","author":"MH Guo","year":"2023","unstructured":"Guo, M.H., Lu, C.Z., Liu, Z.N., Cheng, M.M., Hu, S.M.: Visual attention network, Computational Visual. Media 9(4), 733 (2023)","journal-title":"Media"},{"doi-asserted-by":"crossref","unstructured":"Zou,Z., Chen,K., Shi,Z., Guo, Y., Ye,J.: Object detection in 20 years: A survey, Proceedings of the IEEE (2023)","key":"1549_CR18","DOI":"10.1109\/JPROC.2023.3238524"},{"unstructured":"Ge,Z., Liu,S., Wang,F., Li,Z., Sun,J.: Yolox: Exceeding yolo series in 2021, arXiv preprint arXiv:2107.08430 (2021)","key":"1549_CR19"},{"unstructured":"Redmon,J., Farhadi,A.: Yolov3: An incremental improvement, arXiv preprint arXiv:1804.02767 (2018)","key":"1549_CR20"},{"unstructured":"Jocher,G., Chaurasia,A., Qiu,J: .Jocher, Glenn and Chaurasia, Ayush and Qiu, Jing, Yolo by ultralytics (2023). https:\/\/github.com\/ultralytics\/ultralytics","key":"1549_CR21"},{"doi-asserted-by":"crossref","unstructured":"Chen,Z., Yang,C., Li,Q., Zhao,F., Zha,Z.J., Wu,F.: Disentangle Your Dense Object Detector,TOOD: Task-aligned One-stage Object Detection, In: Proceedings of the 29th ACM International Conference on Multimedia (2021), 4939\u20134948","key":"1549_CR22","DOI":"10.1145\/3474085.3475351"},{"doi-asserted-by":"crossref","unstructured":"Feng,C., Zhong,Y., Gao,Y., Scott,M.R., Huang,W.: TOOD: Task-aligned One-stage Object Detection,in ICCV (2021)","key":"1549_CR23","DOI":"10.1109\/ICCV48922.2021.00349"},{"unstructured":"Liu,S., Li,F., Zhang,H., Yang,X., Qi,X., Su,H., Zhu,J., Zhang,L.: DAB-DETR: Dynamic Anchor Boxes are Better Queries for DETR, in International Conference on Learning Representations (2022). https:\/\/openreview.net\/forum?id=oMI9PjOb9Jl","key":"1549_CR24"},{"unstructured":"Zhang,H., Li,F., Liu,S., Zhang,L., Su,H., Zhu,J., Ni,L.M., Shum,H.Y.: Dino: Detr with improved denoising anchor boxes for end-to-end object detection (2022)","key":"1549_CR25"},{"doi-asserted-by":"crossref","unstructured":"Chen,S., Sun,P., Song,Y., Luo,P.: DiffusionDet: Diffusion Model for Object Detection, arXiv preprint arXiv:2211.09788 (2022)","key":"1549_CR26","DOI":"10.1109\/ICCV51070.2023.01816"},{"doi-asserted-by":"crossref","unstructured":"Zhang,S., Wang,X., Wang,J., Pang,J., Lyu,C., Zhang,W., Luo,P., Chen,K.: Dense Distinct Query for End-to-End Object Detection, In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2023), 7329\u20137338","key":"1549_CR27","DOI":"10.1109\/CVPR52729.2023.00708"},{"doi-asserted-by":"crossref","unstructured":"Zhao,Y., Lv,W., Xu,S., Wei,J., Wang,G., Dang,Q., Liu,Y., Chen,J.: Detrs beat yolos on real-time object detection (2023)","key":"1549_CR28","DOI":"10.1109\/CVPR52733.2024.01605"},{"unstructured":"Bochkovskiy,A., Wang,C.Y., Liao,H.Y.M.:Yolov4: Optimal speed and accuracy of object detection, arXiv preprint arXiv:2004.10934 (2020)","key":"1549_CR29"},{"doi-asserted-by":"publisher","unstructured":"He,K., Zhang,X., Ren,S., Sun,J.: Deep Residual Learning for Image Recognition, In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016), 770\u2013778. https:\/\/doi.org\/10.1109\/CVPR.2016.90","key":"1549_CR30","DOI":"10.1109\/CVPR.2016.90"},{"doi-asserted-by":"crossref","unstructured":"Jocher,G., Chaurasia,A., Stoken,A., Borovec,J., Kwon,Y., Michael,K., Fang,J., Wong,C., Yifu,Z., Montes,D. et\u00a0al.,ultralytics\/yolov5: v6. 2-yolov5 classification models, apple m1, reproducibility, clearml and deci. ai integrations, Zenodo (2022)","key":"1549_CR31","DOI":"10.1109\/IPAS55744.2022.10053056"},{"unstructured":"Li,C., Li,L., Jiang,H., Weng,K., Geng,Y., Li,L., Ke,Z., Li,Q., Cheng,M., Nie,W. et al.,YOLOv6: A single-stage object detection framework for industrial applications, arXiv preprint arXiv:2209.02976 (2022)","key":"1549_CR32"},{"unstructured":"Xu,X., Jiang,Y., Chen,W., Huang,Y., Zhang,Y., Sun,X.:Damo-yolo: A report on real-time object detection design, arXiv preprint arXiv:2211.15444 (2022)","key":"1549_CR33"},{"doi-asserted-by":"crossref","unstructured":"Wang,W., Xie,E., Li,X., Fan,D.P., Song,K., Liang,D., Lu,T., Luo,P., Shao,L.: Pyramid vision transformer: A versatile backbone for dense prediction without convolutions,In: Proceedings of the IEEE\/CVF international conference on computer vision (2021), 568\u2013578","key":"1549_CR34","DOI":"10.1109\/ICCV48922.2021.00061"},{"doi-asserted-by":"crossref","unstructured":"Chen,Y., Dai,X., Chen,D., Liu,M., Dong,X., Yuan,L., Liu,Z.:Mobile-former: Bridging mobilenet and transformer, In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2022), 5270\u20135279","key":"1549_CR35","DOI":"10.1109\/CVPR52688.2022.00520"},{"unstructured":"Yan,H., Li,Z., Li,W., Wang,C., Wu,M., Zhang,C.:Contnet: Why not use convolution and transformer at the same time?, arXiv preprint arXiv:2104.13497 (2021)","key":"1549_CR36"}],"container-title":["Journal of Real-Time Image Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-024-01549-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11554-024-01549-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-024-01549-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,15]],"date-time":"2024-10-15T15:24:00Z","timestamp":1729005840000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11554-024-01549-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,11]]},"references-count":36,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2024,10]]}},"alternative-id":["1549"],"URL":"https:\/\/doi.org\/10.1007\/s11554-024-01549-y","relation":{},"ISSN":["1861-8200","1861-8219"],"issn-type":[{"type":"print","value":"1861-8200"},{"type":"electronic","value":"1861-8219"}],"subject":[],"published":{"date-parts":[[2024,9,11]]},"assertion":[{"value":"11 July 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 August 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 September 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"170"}}