{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T21:39:56Z","timestamp":1773351596295,"version":"3.50.1"},"reference-count":33,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T00:00:00Z","timestamp":1773273600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T00:00:00Z","timestamp":1773273600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100017550","name":"Shaanxi Science and Technology Association","doi-asserted-by":"publisher","award":["2022KXJ- 021"],"award-info":[{"award-number":["2022KXJ- 021"]}],"id":[{"id":"10.13039\/501100017550","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Supercomput"],"DOI":"10.1007\/s11227-026-08395-8","type":"journal-article","created":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T10:59:48Z","timestamp":1773313188000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Scim: a lightweight spatial-channel interaction module for enhancing object detection in autonomous driving"],"prefix":"10.1007","volume":"82","author":[{"given":"Yang","family":"Cui","sequence":"first","affiliation":[]},{"given":"Yi","family":"Han","sequence":"additional","affiliation":[]},{"given":"Dong","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Tian","family":"Guan","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Yang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,3,12]]},"reference":[{"key":"8395_CR1","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.121358","volume":"236","author":"X Zhao","year":"2024","unstructured":"Zhao X, Fang Y, Min H, Xia W, Wang W, Teixeira R (2024) Potential sources of sensor data anomalies for autonomous vehicles: an overview from road vehicle safety perspective. Expert Syst Appl 236:121358","journal-title":"Expert Syst Appl"},{"key":"8395_CR2","doi-asserted-by":"crossref","unstructured":"Srivastav A, Mandal S (2023) Radars for autonomous driving: a review of deep learning methods and challenges. IEEE Access","DOI":"10.1109\/ACCESS.2023.3312382"},{"key":"8395_CR3","doi-asserted-by":"crossref","unstructured":"Ranftl R, Bochkovskiy A, Koltun V (2021) Vision transformers for dense prediction. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 12179\u201312188","DOI":"10.1109\/ICCV48922.2021.01196"},{"key":"8395_CR4","doi-asserted-by":"publisher","DOI":"10.1016\/j.scs.2020.102700","volume":"67","author":"Z Liu","year":"2021","unstructured":"Liu Z, Qi M, Shen C, Fang Y, Zhao X (2021) Cascade saccade machine learning network with hierarchical classes for traffic sign detection. Sustain Cities Soc 67:102700","journal-title":"Sustain Cities Soc"},{"key":"8395_CR5","doi-asserted-by":"publisher","DOI":"10.1016\/j.compag.2022.107062","volume":"198","author":"Y Zhang","year":"2022","unstructured":"Zhang Y, Zhang W, Jiya Yu, He L, Chen J, He Y (2022) Complete and accurate holly fruits counting using YOLOX object detection. Comput Electron Agric 198:107062","journal-title":"Comput Electron Agric"},{"key":"8395_CR6","unstructured":"Dosovitskiy A (2020) An image is worth 16 X 16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929"},{"key":"8395_CR7","unstructured":"Heidari M, Azad R, Kolahi SG, Arimond R, Niggemeier L, Sulaiman A, Bozorgpour A, Aghdam EK, Kazerouni A, Hacihaliloglu I et al (2024) Enhancing efficiency in vision transformer networks: design techniques and insights. arXiv preprint arXiv:2403.19882"},{"key":"8395_CR8","unstructured":"Keles FD, Wijewardena PM, Hegde C (2023) On the computational complexity of self-attention. In: International Conference on Algorithmic Learning Theory, pp 597\u2013619. PMLR"},{"key":"8395_CR9","doi-asserted-by":"publisher","first-page":"5590894","DOI":"10.1155\/2021\/5590894","volume":"1","author":"P Jagannathan","year":"2021","unstructured":"Jagannathan P, Rajkumar S, Frnda J, Divakarachari PB, Subramani P (2021) Moving vehicle detection and classification using gaussian mixture model and ensemble deep learning technique. Wirel Commun Mobile Comput 1:5590894","journal-title":"Wirel Commun Mobile Comput"},{"issue":"9","key":"8395_CR10","doi-asserted-by":"publisher","first-page":"15864","DOI":"10.1109\/TITS.2022.3146087","volume":"23","author":"K Li","year":"2022","unstructured":"Li K, Xiong H, Liu J, Qing X, Wang J (2022) Real-time monocular joint perception network for autonomous driving. IEEE Trans Intell Transp Syst 23(9):15864\u201315877","journal-title":"IEEE Trans Intell Transp Syst"},{"key":"8395_CR11","unstructured":"Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. Adv Neural Inf Process Syst Vol. 25"},{"key":"8395_CR12","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556"},{"key":"8395_CR13","doi-asserted-by":"publisher","first-page":"40701","DOI":"10.1109\/ACCESS.2022.3166923","volume":"10","author":"T Liang","year":"2022","unstructured":"Liang T, Bao H, Pan W, Pan F (2022) ALODAD: an anchor-free lightweight object detector for autonomous driving. IEEE Access 10:40701\u201340714","journal-title":"IEEE Access"},{"key":"8395_CR14","doi-asserted-by":"crossref","unstructured":"Carion N, Massa F, Synnaeve G, Usunier N, Alexander K, Sergey Z (2020) End-to-end object detection with transformers. In: European Conference on Computer Vision, pp 213\u2013229. Springer","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"8395_CR15","unstructured":"Liu S, Li F, Zhang H, Yang X, Qi X, Su H, Zhu J, Zhang L (2021) Dab-detr: dynamic anchor boxes are better queries for DETR. In: International Conference on Learning Representations"},{"key":"8395_CR16","unstructured":"Chen Q, Wang J, Han C, Zhang S, Li Z, Chen X, Chen J, Wang X, Han S, Zhang G et al (2022) Group detr v2: strong object detector with encoder\u2013decoder pretraining. arXiv preprint arXiv:2211.03594"},{"key":"8395_CR17","unstructured":"Zhang C, Liu L, Zang X, Liu F, Zhang H, Song X, Chen J (2022) Detr++: taming your multi-scale detection transformer. arXiv preprint arXiv:2206.02977"},{"key":"8395_CR18","doi-asserted-by":"crossref","unstructured":"Wang W, Xie E, Li X, Fan DP, Song K, Liang D, Lu T, Luo P, Shao L (2021) Pyramid vision transformer: a versatile backbone for dense prediction without convolutions. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 568\u2013578","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"8395_CR19","unstructured":"Touvron H, Cord M, Douze M, Massa F, Sablayrolles A, J\u00e9gou H (2021) Training data-efficient image transformers & distillation through attention. In: International Conference on Machine Learning, pp 10347\u201310357. PMLR"},{"key":"8395_CR20","doi-asserted-by":"crossref","unstructured":"Yuan K, Guo S, Liu Z, Zhou A, Yu F, Wu W (2021) Incorporating convolution designs into visual transformers. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 579\u2013588","DOI":"10.1109\/ICCV48922.2021.00062"},{"key":"8395_CR21","doi-asserted-by":"crossref","unstructured":"Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, Lin S, Guo B (2021) Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 10012\u201310022","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"8395_CR22","doi-asserted-by":"crossref","unstructured":"Liu Z, Mao H, Wu CY, Feichtenhofer C, Trevor D, Saining X (2022) A convnet for the 2020s. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 11976\u201311986.","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"8395_CR23","doi-asserted-by":"crossref","unstructured":"Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 7132\u20137141","DOI":"10.1109\/CVPR.2018.00745"},{"key":"8395_CR24","doi-asserted-by":"crossref","unstructured":"Woo S, Park J, Lee JY, Kweon IS (2018) CBAM: convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp 3\u201319","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"8395_CR25","doi-asserted-by":"crossref","unstructured":"Pan X, Ge C, Lu R, Song S, Chen G, Huang Z, Huang G (2022) On the integration of self-attention and convolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 815\u2013825","DOI":"10.1109\/CVPR52688.2022.00089"},{"key":"8395_CR26","unstructured":"Luo W, Li Y, Urtasun R, Zemel R (2016) Understanding the effective receptive field in deep convolutional neural networks. Adv Neural Inf Process Syst, vol 29"},{"issue":"1","key":"8395_CR27","doi-asserted-by":"publisher","first-page":"16","DOI":"10.1007\/s44267-023-00019-6","volume":"1","author":"D-P Fan","year":"2023","unstructured":"Fan D-P, Ji G-P, Peng X, Cheng M-M, Sakaridis C, Van Gool L (2023) Advances in deep concealed scene understanding. Vis Intell 1(1):16","journal-title":"Vis Intell"},{"key":"8395_CR28","doi-asserted-by":"crossref","unstructured":"Ding X, Zhang X, Han J, Ding G (2022) Scaling up your kernels to 31 x 31: revisiting large kernel design in CNNs. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 11963\u201311975","DOI":"10.1109\/CVPR52688.2022.01166"},{"key":"8395_CR29","doi-asserted-by":"crossref","unstructured":"Wang A, Chen H, Liu L, Chen K, Lin Z, Han J, Ding G (2024) Yolov10: real-time end-to-end object detection. arXiv preprint arXiv:2405.14458","DOI":"10.52202\/079017-3429"},{"key":"8395_CR30","unstructured":"Glenn J (2022) YOLOv5 release v7.0. https:\/\/github.com\/ultralytics\/yolov5\/tree\/v7.0"},{"key":"8395_CR31","unstructured":"Li C, Li L, Geng Y, Jiang H, Cheng M, Zhang B, Ke Z, Xu X, Chu X (2023) YOLOv6 v3.0: a full-scale reloading. arXiv preprint arXiv:2301.05586"},{"key":"8395_CR32","doi-asserted-by":"crossref","unstructured":"Zhao Y, Lv W, Xu S, Wei J, Wang G, Dang Q, Liu Y, Chen J (2024) DETRs beat YOLOs on real-time object detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 16965\u201316974","DOI":"10.1109\/CVPR52733.2024.01605"},{"key":"8395_CR33","doi-asserted-by":"crossref","unstructured":"Yu F, Chen H, Wang X, Xian W, Chen Y, Liu F, Madhavan V, Darrell T (2020) Bdd100k: a diverse driving dataset for heterogeneous multitask learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 2636\u20132645","DOI":"10.1109\/CVPR42600.2020.00271"}],"container-title":["The Journal of Supercomputing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11227-026-08395-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11227-026-08395-8","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11227-026-08395-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T11:00:29Z","timestamp":1773313229000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11227-026-08395-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,3,12]]},"references-count":33,"journal-issue":{"issue":"4","published-online":{"date-parts":[[2026,3]]}},"alternative-id":["8395"],"URL":"https:\/\/doi.org\/10.1007\/s11227-026-08395-8","relation":{},"ISSN":["1573-0484"],"issn-type":[{"value":"1573-0484","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,3,12]]},"assertion":[{"value":"5 November 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 February 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 March 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing financial interests or personal relationships that could have influenced the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"246"}}