{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,7]],"date-time":"2026-02-07T11:34:05Z","timestamp":1770464045398,"version":"3.49.0"},"reference-count":40,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,12,27]],"date-time":"2025-12-27T00:00:00Z","timestamp":1766793600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,12,27]],"date-time":"2025-12-27T00:00:00Z","timestamp":1766793600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Science and Technology Innovation Team in College of Hunan Province, China","award":["2023-233 [XiangJiaoTong]"],"award-info":[{"award-number":["2023-233 [XiangJiaoTong]"]}]},{"name":"Science and Technology Innovation Team in College of Hunan Province, China","award":["2023-233 [XiangJiaoTong]"],"award-info":[{"award-number":["2023-233 [XiangJiaoTong]"]}]},{"name":"Science and Technology Innovation Team in College of Hunan Province, China","award":["2023-233 [XiangJiaoTong]"],"award-info":[{"award-number":["2023-233 [XiangJiaoTong]"]}]},{"name":"Science and Technology Innovation Team in College of Hunan Province, China","award":["2023-233 [XiangJiaoTong]"],"award-info":[{"award-number":["2023-233 [XiangJiaoTong]"]}]},{"name":"Science and Technology Innovation Team in College of Hunan Province, China","award":["2023-233 [XiangJiaoTong]"],"award-info":[{"award-number":["2023-233 [XiangJiaoTong]"]}]},{"name":"Key Laboratory in College of Hunan Province, China","award":["No.2023-213 [XiangJiaoTong]"],"award-info":[{"award-number":["No.2023-213 [XiangJiaoTong]"]}]},{"name":"Key Laboratory in College of Hunan Province, China","award":["No.2023-213 [XiangJiaoTong]"],"award-info":[{"award-number":["No.2023-213 [XiangJiaoTong]"]}]},{"name":"Key Laboratory in College of Hunan Province, China","award":["No.2023-213 [XiangJiaoTong]"],"award-info":[{"award-number":["No.2023-213 [XiangJiaoTong]"]}]},{"name":"Key Laboratory in College of Hunan Province, China","award":["No.2023-213 [XiangJiaoTong]"],"award-info":[{"award-number":["No.2023-213 [XiangJiaoTong]"]}]},{"name":"Key Laboratory in College of Hunan Province, China","award":["No.2023-213 [XiangJiaoTong]"],"award-info":[{"award-number":["No.2023-213 [XiangJiaoTong]"]}]},{"name":"Young Talent Program of Hunan Province","award":["2023RC3197"],"award-info":[{"award-number":["2023RC3197"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Real-Time Image Proc"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1007\/s11554-025-01838-0","type":"journal-article","created":{"date-parts":[[2025,12,27]],"date-time":"2025-12-27T02:03:13Z","timestamp":1766800993000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["LDR-YOLO: lightweight dynamic upsampling and regional attention enhancement network for steel surface defect detection"],"prefix":"10.1007","volume":"23","author":[{"given":"Shanshan","family":"Tan","sequence":"first","affiliation":[]},{"given":"Jinwen","family":"Tan","sequence":"additional","affiliation":[]},{"given":"Junfeng","family":"Man","sequence":"additional","affiliation":[]},{"given":"Junjie","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Xiaoxue","family":"Ding","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,12,27]]},"reference":[{"key":"1838_CR1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.02617","author":"X Cai","year":"2024","unstructured":"Cai, X., Lai, Q., Wang, Y., et al.: Poly kernel inception network for remote sensing detection. IEEE (2024). https:\/\/doi.org\/10.1109\/CVPR52733.2024.02617","journal-title":"IEEE"},{"key":"1838_CR2","doi-asserted-by":"publisher","first-page":"31941","DOI":"10.1038\/s41598-024-83430-3","volume":"14","author":"S Chan","year":"2024","unstructured":"Chan, S., Li, S., Zhang, H., et al.: Feature optimization-guided high-precision and real-time metal surface defect detection network. Sci Rep 14, 31941 (2024). https:\/\/doi.org\/10.1038\/s41598-024-83430-3","journal-title":"Sci Rep"},{"key":"1838_CR3","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01157","author":"J Chen","year":"2023","unstructured":"Chen, J., Kao, S.H., He, H., et al.: Run, don't walk: chasing higher FLOPS for faster neural networks. IEEE (2023). https:\/\/doi.org\/10.1109\/CVPR52729.2023.01157","journal-title":"IEEE"},{"key":"1838_CR4","doi-asserted-by":"crossref","unstructured":"Chen, J., Kao, S-h., He, H., Zhuo, W., Wen, S., Lee, C-H., Gary\u00a0Chan, S.-H.: Run, don\u2019t walk: chasing higher flops for faster neural networks. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12021\u201312031 (2023)","DOI":"10.1109\/CVPR52729.2023.01157"},{"issue":"12","key":"1838_CR5","doi-asserted-by":"publisher","first-page":"20103","DOI":"10.1109\/JSEN.2024.3392674","volume":"24","author":"X Chen","year":"2024","unstructured":"Chen, X., Jiang, Z., Piao, Y., Yang, J., Zheng, H., Yang, H., Chen, K.: SF-Yolov8n: a novel ultralightweight and high-precision model for detecting surface defects of dental nails. IEEE Sens. J. 24(12), 20103\u201320113 (2024)","journal-title":"IEEE Sens. J."},{"key":"1838_CR6","doi-asserted-by":"publisher","DOI":"10.1007\/s11554-025-01630-0","author":"W Cheng","year":"2025","unstructured":"Cheng, W., Zeng, P., Hao, Y.: Lightweight YOLOv7 for bushing surface defects detection. Journal of Real-Time Image Processing (2025). https:\/\/doi.org\/10.1007\/s11554-025-01630-0","journal-title":"Journal of Real-Time Image Processing"},{"key":"1838_CR7","doi-asserted-by":"crossref","unstructured":"Chollet, F.: Xception: deep learning with depthwise separable convolutions. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1800\u20131807 (2017)","DOI":"10.1109\/CVPR.2017.195"},{"key":"1838_CR8","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.195","author":"F Chollet","year":"2017","unstructured":"Chollet, F.: Xception: Deep Learning with Depthwise Separable Convolutions. IEEE (2017). https:\/\/doi.org\/10.1109\/CVPR.2017.195","journal-title":"IEEE"},{"key":"1838_CR9","doi-asserted-by":"publisher","unstructured":"Duan K , Bai S , Xie L ,et al. CenterNet: Keypoint Triplets for Object Detection. 2019. https:\/\/doi.org\/10.48550\/arXiv.1904.08189","DOI":"10.48550\/arXiv.1904.08189"},{"key":"1838_CR10","doi-asserted-by":"crossref","unstructured":"Girshick, R.: Fast R-CNN. In: 2015 IEEE International Conference on Computer Vision (ICCV), pp. 1440\u20131448 (2015)","DOI":"10.1109\/ICCV.2015.169"},{"issue":"000","key":"1838_CR11","doi-asserted-by":"publisher","first-page":"48188","DOI":"10.1109\/ACCESS.2025.3550374","volume":"13","author":"W Guiqiang","year":"2025","unstructured":"Guiqiang, W., Junbao, C., Chengzhang, L., et al.:  Edge-YOLO: Lightweight multi-Scale feature extraction for industrial surface inspection. Access, IEEE 13(000), 48188\u201348201 (2025). https:\/\/doi.org\/10.1109\/ACCESS.2025.3550374","journal-title":"Access, IEEE"},{"key":"1838_CR12","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00165","author":"K Han","year":"2020","unstructured":"Han, K., Wang, Y., Tian, Q., et al.: GhostNet: more features from cheap operations. IEEE (2020). https:\/\/doi.org\/10.1109\/CVPR42600.2020.00165","journal-title":"IEEE"},{"key":"1838_CR13","doi-asserted-by":"publisher","DOI":"10.1016\/j.dsp.2024.104615","volume":"153","author":"Y Han","year":"2024","unstructured":"Han, Y., Wang, F., Wang, W., Zhang, X., Li, X.: EDN-YOLO: multi-scale traffic sign detection method in complex scenes. Digit. Signal Process. 153, 104615 (2024)","journal-title":"Digit. Signal Process."},{"key":"1838_CR14","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. IEEE (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"1838_CR15","doi-asserted-by":"crossref","unstructured":"Howard, A., Sandler, M., Chen, B., Wang, W., Chen, L.C.., Tan, M., Chu, G., Vasudevan, V., Zhu, Y., Pang, R.: Searching for MobileNetV3. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV) (2020)","DOI":"10.1109\/ICCV.2019.00140"},{"key":"1838_CR16","doi-asserted-by":"publisher","unstructured":"Howard A G , Zhu M , Chen B, et al. MobileNets: efficient convolutional neural networks for mobile vision applications. 2017. https:\/\/doi.org\/10.48550\/arXiv.1704.04861DOI:.","DOI":"10.48550\/arXiv.1704.04861"},{"key":"1838_CR17","doi-asserted-by":"crossref","unstructured":"Li, Y., Hou, Q., Zheng, Z., Cheng, M.-M., Yang, J., Li, X.: Large selective kernel network for remote sensing object detection. In: 2023 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 16748\u201316759 (2023)","DOI":"10.1109\/ICCV51070.2023.01540"},{"key":"1838_CR18","doi-asserted-by":"crossref","unstructured":"Liu, W., Lu, H., Fu, H., Cao, Z.: Learning to upsample by learning to sample. In: 2023 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 6004\u20136014 (2023)","DOI":"10.1109\/ICCV51070.2023.00554"},{"key":"1838_CR19","doi-asserted-by":"publisher","DOI":"10.3390\/s20061562","author":"X Lv","year":"2020","unstructured":"Lv, X., Duan, F., Jiang, J.J., et al.: Deep metallic surface defect detection: the new benchmark and detection network. Sensors (2020). https:\/\/doi.org\/10.3390\/s20061562","journal-title":"Sensors"},{"issue":"9","key":"1838_CR20","doi-asserted-by":"publisher","first-page":"6870","DOI":"10.1109\/TIM.2020.2975454","volume":"69","author":"M Nieniewski","year":"2020","unstructured":"Nieniewski, M.: Morphological detection and extraction of rail surface defects. IEEE Trans. Instrum. Meas. 69(9), 6870\u20136879 (2020)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"1838_CR21","doi-asserted-by":"publisher","unstructured":"Qin D , Leichner C , Delakis M ,et al.MobileNetV4: Universal Models fortheMobile Ecosystem[C]\/\/European Conference on Computer Vision.Springer, Cham, 2025.https:\/\/doi.org\/10.1007\/978-3-031-73661-2_5","DOI":"10.1007\/978-3-031-73661-2_5"},{"key":"1838_CR22","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779\u2013788 (2016)","DOI":"10.1109\/CVPR.2016.91"},{"issue":"6","key":"1838_CR23","doi-asserted-by":"publisher","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","volume":"39","author":"S Ren","year":"2017","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(6), 1137\u20131149 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"3","key":"1838_CR24","doi-asserted-by":"publisher","first-page":"2183","DOI":"10.1007\/s11760-023-02884-6","volume":"18","author":"G Revathy","year":"2024","unstructured":"Revathy, G., Kalaivani, R.: Fabric defect detection and classification via deep learning-based improved mask RCNN. Signal Image Video Process. 18(3), 2183\u20132193 (2024)","journal-title":"Signal Image Video Process."},{"key":"1838_CR25","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.-C.: MobileNetV2: inverted residuals and linear bottlenecks. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"1838_CR26","doi-asserted-by":"publisher","unstructured":"Sifre L ,Mallat, St\u00e9phane.Rigid-Motion Scattering for Texture Classification[J].Computer Science, 2014, 3559:501-515. https:\/\/doi.org\/10.1007\/11503415_34","DOI":"10.1007\/11503415_34"},{"key":"1838_CR27","doi-asserted-by":"publisher","unstructured":"Simonyan K , Zisserman A .Very deep convolutional networks for large-scale image recognition. Comput Sci, 2014. https:\/\/doi.org\/10.48550\/arXiv.1409.1556","DOI":"10.48550\/arXiv.1409.1556"},{"key":"1838_CR28","doi-asserted-by":"publisher","unstructured":"Tan M , Le Q V .EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. 2019. https:\/\/doi.org\/10.48550\/arXiv.1905.11946","DOI":"10.48550\/arXiv.1905.11946"},{"key":"1838_CR29","unstructured":"Tang, Y., Han, K., Guo, J., Xu, C., Xu, C., Wang, Y.: GhostNetV2: enhance cheap operation with long-range attention. ArXiv. arXiv:2211.12905 (2022)"},{"key":"1838_CR30","unstructured":"Tian Y , Ye Q , Doermann D .YOLOv12: Attention-Centric Real-Time Object Detectors[J]. 2025. arXiv:2502.12524"},{"key":"1838_CR31","doi-asserted-by":"crossref","unstructured":"Tiwari, S., Sharma, S.K.: YOLOv7-tiny-based lightweight and efficient algorithm for photovoltaic cell crack detection. J. Supercomput. 81(8) (2025)","DOI":"10.1007\/s11227-025-07493-3"},{"key":"1838_CR32","doi-asserted-by":"publisher","unstructured":"Vaswani A , Shazeer N , Parmar N ,et al.Attention Is All You Need.arXiv, 2017. https:\/\/doi.org\/10.48550\/arXiv.1706.03762.","DOI":"10.48550\/arXiv.1706.03762."},{"key":"1838_CR33","unstructured":"Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., Han, J., Ding, G.: YOLOv10: real-time end-to-end object detection (2024). arXiv:2405.14458"},{"key":"1838_CR34","doi-asserted-by":"publisher","unstructured":"Wang Y , Wang Z , Liu W ,et al.A Novel Depth-Connected Region-Based Convolutional Neural Network for Small Defect Detection in Additive Manufacturing. Cognitive Computation, 2025, 17(1). https:\/\/doi.org\/10.1007\/s12559-024-10397-8DOI:","DOI":"10.1007\/s12559-024-10397-8"},{"key":"1838_CR35","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46448-0_2","volume-title":"SSD: single shot multibox detector","author":"L Wei","year":"2016","unstructured":"Wei L , Dragomir A , Dumitru E ,et al.SSD: Single Shot MultiBox Detector. Springer, Cham, 2016. https:\/\/doi.org\/10.1007\/978-3-319-46448-0_2"},{"issue":"13","key":"1838_CR36","doi-asserted-by":"publisher","first-page":"7181","DOI":"10.1109\/JSEN.2020.2977366","volume":"20","author":"L Xiao","year":"2020","unstructured":"Xiao L , Wu B , Hu Y .Surface Defect Detection Using Image Pyramid[J].IEEE Sensors Journal, 2020, PP(99):1\u20131. https:\/\/doi.org\/10.1109\/JSEN.2020.2977366","journal-title":"IEEE Sens. J."},{"key":"1838_CR37","doi-asserted-by":"crossref","unstructured":"Yan, S., Chen, G., Gao, A., Liu, C., Xiong, Z.: BiSPD-YOLO: surface defect detection method for small features and low-resolution images. In: 2023 IEEE\/ASME International Conference on Advanced Intelligent Mechatronics (AIM), pp. 709\u2013714 (2023)","DOI":"10.1109\/AIM46323.2023.10196194"},{"key":"1838_CR38","doi-asserted-by":"crossref","unstructured":"Zhang, D., Hao, X., Wang, D., Qin, C., Zhao, B., Liang, L., Liu, W.: An efficient lightweight convolutional neural network for industrial surface defect detection. Artif. Intell. Rev. Int. Sci. Eng. J. 56(9) (2023)","DOI":"10.1007\/s10462-023-10438-y"},{"key":"1838_CR39","doi-asserted-by":"crossref","unstructured":"Zhao, Y., Lv, W., Xu, S., Wei, J., Wang, G., Dang, Q., Liu, Y., Chen, J.: DETRs beat YOLOs on real-time object detection. In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 16965\u201316974 (2024)","DOI":"10.1109\/CVPR52733.2024.01605"},{"issue":"6","key":"1838_CR40","doi-asserted-by":"publisher","first-page":"9197","DOI":"10.1109\/JSEN.2024.3359218","volume":"24","author":"L Zuo","year":"2024","unstructured":"L. Zuo et al., \"A New Dual-Branch Network With Global Information for the Surface Defect Detection on Solar PV Wafer,\" in IEEE Sensors Journal, vol. 24, no. 6, pp. 9197-9207, 15 March15, 2024, doi: 10.1109\/JSEN.2024.3359218. keywords: {Feature extraction;Inspection;Wheels;Surface morphology;Steel;Sensors;Manufacturing;Attention mechanism;global information;multiscale feature fusion;surface defect detection (SDD)},","journal-title":"IEEE Sens. J."}],"container-title":["Journal of Real-Time Image Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-025-01838-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11554-025-01838-0","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-025-01838-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T16:49:53Z","timestamp":1770396593000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11554-025-01838-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,27]]},"references-count":40,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,2]]}},"alternative-id":["1838"],"URL":"https:\/\/doi.org\/10.1007\/s11554-025-01838-0","relation":{},"ISSN":["1861-8200","1861-8219"],"issn-type":[{"value":"1861-8200","type":"print"},{"value":"1861-8219","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,27]]},"assertion":[{"value":"27 August 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 December 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 December 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"41"}}