{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,7]],"date-time":"2026-02-07T11:34:11Z","timestamp":1770464051773,"version":"3.49.0"},"reference-count":38,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,12,16]],"date-time":"2025-12-16T00:00:00Z","timestamp":1765843200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,12,16]],"date-time":"2025-12-16T00:00:00Z","timestamp":1765843200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"the Fujian Provincial Academy of Sciences STS Project","award":["2023T3030"],"award-info":[{"award-number":["2023T3030"]}]},{"name":"the Fujian Provincial Academy of Sciences STS Project","award":["2023T3030"],"award-info":[{"award-number":["2023T3030"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Real-Time Image Proc"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1007\/s11554-025-01828-2","type":"journal-article","created":{"date-parts":[[2025,12,16]],"date-time":"2025-12-16T09:31:53Z","timestamp":1765877513000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Real-time safety detection based on dynamic multi-scale feature fusion for forklift driving"],"prefix":"10.1007","volume":"23","author":[{"given":"Songhang","family":"Chen","sequence":"first","affiliation":[]},{"given":"Jiang","family":"Wu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,12,16]]},"reference":[{"key":"1828_CR1","unstructured":"Bochkovskiy, A., Wang, C.Y., Liao, H.Y.M.: YOLOv4: optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934 (2020)"},{"key":"1828_CR2","doi-asserted-by":"crossref","unstructured":"Chen, J., Kao, S.H., He, H., Zhuo, W., Wen, S., Lee, C.H., Chan, S.H.G.: Run, don\u2019t walk: chasing higher flops for faster neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12021\u201312031 (2023)","DOI":"10.1109\/CVPR52729.2023.01157"},{"key":"1828_CR3","doi-asserted-by":"crossref","unstructured":"Dai, J., Qi, H., Xiong, Y., Li, Y., Zhang, G., Hu, H., Wei, Y.: Deformable convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 764\u2013773 (2017)","DOI":"10.1109\/ICCV.2017.89"},{"key":"1828_CR4","doi-asserted-by":"crossref","unstructured":"Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. In: 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR\u201905), vol.\u00a01, pp. 886\u2013893. IEEE (2005)","DOI":"10.1109\/CVPR.2005.177"},{"key":"1828_CR5","doi-asserted-by":"crossref","unstructured":"Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving the KITTI vision benchmark suite. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 3354\u20133361. IEEE (2012)","DOI":"10.1109\/CVPR.2012.6248074"},{"key":"1828_CR6","unstructured":"Gevorgyan, Z.: SIoU loss: more powerful learning for bounding box regression. arXiv preprint arXiv:2205.12740 (2022)"},{"key":"1828_CR7","doi-asserted-by":"crossref","unstructured":"Girshick, R.: Fast R-CNN. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1440\u20131448 (2015)","DOI":"10.1109\/ICCV.2015.169"},{"key":"1828_CR8","doi-asserted-by":"crossref","unstructured":"Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 580\u2013587 (2014)","DOI":"10.1109\/CVPR.2014.81"},{"key":"1828_CR9","doi-asserted-by":"crossref","unstructured":"Han, K., Wang, Y., Tian, Q., Guo, J., Xu, C., Xu, C.: GhostNet: more features from cheap operations. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1580\u20131589 (2020)","DOI":"10.1109\/CVPR42600.2020.00165"},{"key":"1828_CR10","unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., Adam, H.: MobileNets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017)"},{"issue":"3","key":"1828_CR11","doi-asserted-by":"publisher","first-page":"1010","DOI":"10.1109\/TITS.2018.2838132","volume":"20","author":"X Hu","year":"2018","unstructured":"Hu, X., Xu, X., Xiao, Y., Chen, H., He, S., Qin, J., Heng, P.A.: SINet: a scale-insensitive convolutional neural network for fast vehicle detection. IEEE Trans. Intell. Transp. Syst. 20(3), 1010\u20131019 (2018)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"1828_CR12","unstructured":"Iandola, F.N., Han, S., Moskewicz, M.W., Ashraf, K., Dally, W.J., Keutzer, K.: SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and $$<$$ 0.5 MB model size. arXiv preprint arXiv:1602.07360 (2016)"},{"key":"1828_CR13","first-page":"1","volume":"73","author":"L Jiang","year":"2024","unstructured":"Jiang, L., Yuan, B., Du, J., Chen, B., Xie, H., Tian, J., Yuan, Z.: MFFSODNet: multi-scale feature fusion small object detection network for UAV aerial images. IEEE Trans. Instrum. Meas. 73, 1\u201314 (2024)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"1828_CR14","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Goyal, P., Girshick, R., He, K., Doll\u00e1r, P.: Focal loss for dense object detection. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2980\u20132988 (2017)","DOI":"10.1109\/ICCV.2017.324"},{"key":"1828_CR15","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., Berg, A.C.: SSD: single shot multibox detector. In: Computer Vision\u2013ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, Part I 14, pp. 21\u201337. Springer (2016)","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"1828_CR16","doi-asserted-by":"crossref","unstructured":"Liu, W., Lu, H., Fu, H., Cao, Z.: Learning to upsample by learning to sample. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6027\u20136037 (2023)","DOI":"10.1109\/ICCV51070.2023.00554"},{"key":"1828_CR17","doi-asserted-by":"crossref","unstructured":"Ma, N., Zhang, X., Zheng, H.T., Sun, J.: ShuffleNet v2: practical guidelines for efficient CNN architecture design. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 116\u2013131 (2018)","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"1828_CR18","unstructured":"Ma, S., Xu, Y.: MPDIoU: a loss for efficient and accurate bounding box regression. arXiv preprint arXiv:2307.07662 (2023)"},{"key":"1828_CR19","doi-asserted-by":"publisher","first-page":"55389","DOI":"10.1109\/ACCESS.2023.3282110","volume":"11","author":"TP Nguyen","year":"2023","unstructured":"Nguyen, T.P., Nam, H., Kim, D.: Transformer-based attention network for in-vehicle intrusion detection. IEEE Access 11, 55389\u201355403 (2023)","journal-title":"IEEE Access"},{"key":"1828_CR20","doi-asserted-by":"crossref","unstructured":"Ouyang, D., He, S., Zhang, G., Luo, M., Guo, H., Zhan, J., Huang, Z.: Efficient multi-scale attention module with cross-spatial learning. In: ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1\u20135. IEEE (2023)","DOI":"10.1109\/ICASSP49357.2023.10096516"},{"key":"1828_CR21","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. In: Advances in Neural Information Processing Systems, vol. 28 (2015)"},{"key":"1828_CR22","doi-asserted-by":"crossref","unstructured":"Rezatofighi, H., Tsoi, N., Gwak, J., Sadeghian, A., Reid, I., Savarese, S.: Generalized intersection over union: a metric and a loss for bounding box regression. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 658\u2013666 (2019)","DOI":"10.1109\/CVPR.2019.00075"},{"key":"1828_CR23","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.C.: MobileNetv2: inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"1828_CR24","unstructured":"Tong, Z., Chen, Y., Xu, Z., Yu, R.: Wise-IoU: bounding box regression loss with dynamic focusing mechanism. arXiv preprint arXiv:2301.10051 (2023)"},{"key":"1828_CR25","first-page":"107984","volume":"37","author":"A Wang","year":"2024","unstructured":"Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., Han, J., et al.: YOLOv10: real-time end-to-end object detection. Adv. Neural Inf. Process. Syst. 37, 107984\u2013108011 (2024)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"1828_CR26","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Bochkovskiy, A., Liao, H.Y.M.: YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7464\u20137475 (2023)","DOI":"10.1109\/CVPR52729.2023.00721"},{"key":"1828_CR27","first-page":"1","volume":"62","author":"J Xu","year":"2024","unstructured":"Xu, J., Fan, X., Jian, H., Xu, C., Bei, W., Ge, Q., Zhao, T.: YoloOW: a spatial scale adaptive real-time object detection neural network for open water search and rescue from UAV aerial imagery. IEEE Trans. Geosci. Remote Sens. 62, 1\u201315 (2024)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"1828_CR28","first-page":"1","volume":"62","author":"H Yang","year":"2024","unstructured":"Yang, H., Liu, J., Wang, Z., Fu, Z., Tan, Q., Niu, S.: MAPFF: multi-angle pyramid feature fusion network for infrared dim small target detection. IEEE Trans. Geosci. Remote Sens. 62, 1\u201316 (2024)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"1828_CR29","doi-asserted-by":"publisher","first-page":"13887","DOI":"10.1109\/TITS.2024.3390576","volume":"25","author":"Z Yao","year":"2024","unstructured":"Yao, Z., Liu, Q., Fu, J., Xie, Q., Li, B., Ye, Q., Li, Q.: A coarse-to-fine deep learning based framework for traffic light recognition. IEEE Trans. Intell. Transp. Syst. 25, 13887\u201313899 (2024)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"issue":"9","key":"1828_CR30","doi-asserted-by":"publisher","first-page":"9736","DOI":"10.1109\/TITS.2023.3267430","volume":"24","author":"Z Yao","year":"2023","unstructured":"Yao, Z., Liu, Q., Xie, Q., Li, Q.: TL-Detector: lightweight based real-time traffic light detection model for intelligent vehicles. IEEE Trans. Intell. Transp. Syst. 24(9), 9736\u20139750 (2023)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"1828_CR31","doi-asserted-by":"publisher","first-page":"4410","DOI":"10.1109\/TSMC.2025.3552621","volume":"55","author":"Z Yao","year":"2025","unstructured":"Yao, Z., Liu, Q., Zhao, Z., Qin, Y., Zhu, J., Xia, T., Li, B., Wang, L.: Night-time traffic light recognition based on enhancement-guided object detection. IEEE Trans. Syst. Man Cybern. Syst. 55, 4410\u20134422 (2025)","journal-title":"IEEE Trans. Syst. Man Cybern. Syst."},{"key":"1828_CR32","unstructured":"Zhang, H., Zhang, S.: Shape-IoU: more accurate metric considering bounding box shape and scale. arXiv preprint arXiv:2312.17663 (2023)"},{"key":"1828_CR33","doi-asserted-by":"crossref","unstructured":"Zhang, X., Zhou, X., Lin, M., Sun, J.: ShuffleNet: an extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6848\u20136856 (2018)","DOI":"10.1109\/CVPR.2018.00716"},{"key":"1828_CR34","first-page":"1","volume":"62","author":"Y Zhang","year":"2024","unstructured":"Zhang, Y., Ye, M., Zhu, G., Liu, Y., Guo, P., Yan, J.: FFCA-YOLO for small object detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 62, 1\u201315 (2024)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"1828_CR35","doi-asserted-by":"publisher","first-page":"146","DOI":"10.1016\/j.neucom.2022.07.042","volume":"506","author":"YF Zhang","year":"2022","unstructured":"Zhang, Y.F., Ren, W., Zhang, Z., Jia, Z., Wang, L., Tan, T.: Focal and efficient IoU loss for accurate bounding box regression. Neurocomputing 506, 146\u2013157 (2022)","journal-title":"Neurocomputing"},{"key":"1828_CR36","doi-asserted-by":"publisher","first-page":"8590","DOI":"10.1109\/ACCESS.2022.3143365","volume":"10","author":"J Zhao","year":"2022","unstructured":"Zhao, J., Hao, S., Dai, C., Zhang, H., Zhao, L., Ji, Z., Ganchev, I.: Improved vision-based vehicle detection and classification by optimized YOLOv4. IEEE Access 10, 8590\u20138603 (2022)","journal-title":"IEEE Access"},{"key":"1828_CR37","doi-asserted-by":"crossref","unstructured":"Zheng, Z., Wang, P., Liu, W., Li, J., Ye, R., Ren, D.: Distance-IoU loss: faster and better learning for bounding box regression. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a034, pp. 12993\u201313000 (2020)","DOI":"10.1609\/aaai.v34i07.6999"},{"key":"1828_CR38","doi-asserted-by":"crossref","unstructured":"Zitnick, C.L., Doll\u00e1r, P.: Edge boxes: locating object proposals from edges. In: Computer Vision\u2013ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6\u201312, 2014, Proceedings, Part V 13, pp. 391\u2013405. Springer (2014)","DOI":"10.1007\/978-3-319-10602-1_26"}],"container-title":["Journal of Real-Time Image Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-025-01828-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11554-025-01828-2","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-025-01828-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T16:50:09Z","timestamp":1770396609000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11554-025-01828-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,16]]},"references-count":38,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,2]]}},"alternative-id":["1828"],"URL":"https:\/\/doi.org\/10.1007\/s11554-025-01828-2","relation":{},"ISSN":["1861-8200","1861-8219"],"issn-type":[{"value":"1861-8200","type":"print"},{"value":"1861-8219","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,16]]},"assertion":[{"value":"11 April 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 December 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 December 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"30"}}