{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,3]],"date-time":"2025-12-03T01:00:42Z","timestamp":1764723642939,"version":"3.46.0"},"reference-count":35,"publisher":"Springer Science and Business Media LLC","issue":"16","license":[{"start":{"date-parts":[[2025,11,10]],"date-time":"2025-11-10T00:00:00Z","timestamp":1762732800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,11,10]],"date-time":"2025-11-10T00:00:00Z","timestamp":1762732800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Shanghai Educational Science Research General Project","award":["C2022120","C2022120","C2022120"],"award-info":[{"award-number":["C2022120","C2022120","C2022120"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s11760-025-04924-9","type":"journal-article","created":{"date-parts":[[2025,11,10]],"date-time":"2025-11-10T00:45:14Z","timestamp":1762735514000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["EA-DETR: an enhanced attention and multi-scale fusion detector for UAV images"],"prefix":"10.1007","volume":"19","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-8054-302X","authenticated-orcid":false,"given":"Qi","family":"Cao","sequence":"first","affiliation":[]},{"given":"Zhigang","family":"Lian","sequence":"additional","affiliation":[]},{"given":"Yadi","family":"Zhao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,11,10]]},"reference":[{"key":"4924_CR1","volume":"42","author":"C Huang","year":"2024","unstructured":"Huang, C., Fang, S., Wu, H., Wang, Y., Yang, Y.: Low-altitude intelligent transportation: system architecture, infrastructure, and key technologies. J. Ind. Inf. Integr. 42, 100694 (2024)","journal-title":"J. Ind. Inf. Integr."},{"key":"4924_CR2","doi-asserted-by":"crossref","unstructured":"Zhou, Y.: Unmanned aerial vehicles based low-altitude economy with lifecycle techno-economic-environmental analysis for sustainable and smart cities. J. Clean. Prod. or J Clean Prod., 145050 (2025)","DOI":"10.1016\/j.jclepro.2025.145050"},{"key":"4924_CR3","doi-asserted-by":"crossref","unstructured":"Girshick, R.: Fast r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1440\u20131448 (2015)","DOI":"10.1109\/ICCV.2015.169"},{"key":"4924_CR4","unstructured":"Jocher, G., Stoken, A., Borovec, J., Changyu, L., Hogan, A., Diaconu, L., Poznanski, J., Yu, L., Rai, P., Ferriday, R., et al.: ultralytics\/yolov5: v3. 0. Zenodo (2020)"},{"key":"4924_CR5","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: Unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779\u2013788 (2016)","DOI":"10.1109\/CVPR.2016.91"},{"key":"4924_CR6","doi-asserted-by":"crossref","unstructured":"Wang, C.-Y., Bochkovskiy, A., Liao, H.-Y.M.: Yolov7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7464\u20137475 (2023)","DOI":"10.1109\/CVPR52729.2023.00721"},{"key":"4924_CR7","unstructured":"Ge, Z., Liu, S., Wang, F., Li, Z., Sun, J.: Yolox: Exceeding yolo series in 2021. arXiv preprint arXiv:2107.08430 (2021)"},{"key":"4924_CR8","unstructured":"Jocher, G., Chaurasia, A., Qiu, J.: Ultralytics YOLOv8. https:\/\/github.com\/ultralytics\/ultralytics"},{"key":"4924_CR9","doi-asserted-by":"crossref","unstructured":"Wang, C.-Y., Yeh, I.-H., Mark\u00a0Liao, H.-Y.: Yolov9: Learning what you want to learn using programmable gradient information. In: European Conference on Computer Vision, pp. 1\u201321 (2024). Springer","DOI":"10.1007\/978-3-031-72751-1_1"},{"key":"4924_CR10","first-page":"107984","volume":"37","author":"A Wang","year":"2024","unstructured":"Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., Han, J., et al.: Yolov10: Real-time end-to-end object detection. Adv. Neural. Inf. Process. Syst. 37, 107984\u2013108011 (2024)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"4924_CR11","unstructured":"Khanam, R., Hussain, M.: Yolov11: An overview of the key architectural enhancements. arXiv:2410.17725 (2024)"},{"key":"4924_CR12","unstructured":"Shehzadi, T., Hashmi, K.A., Stricker, D., Afzal, M.Z.: Object detection with transformers: A review. arXiv:2306.04670 (2023)"},{"key":"4924_CR13","doi-asserted-by":"crossref","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: European Conference on Computer Vision, pp. 213\u2013229 (2020). Springer","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"4924_CR14","unstructured":"Zhu, X., Su, W., Lu, L., Li, B., Wang, X., Dai, J.: Deformable detr: Deformable transformers for end-to-end object detection. arXiv:2010.04159 (2020)"},{"key":"4924_CR15","doi-asserted-by":"crossref","unstructured":"Zhao, Y., Lv, W., Xu, S., Wei, J., Wang, G., Dang, Q., Liu, Y., Chen, J.: Detrs beat yolos on real-time object detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16965\u201316974 (2024)","DOI":"10.1109\/CVPR52733.2024.01605"},{"issue":"10","key":"4924_CR16","doi-asserted-by":"publisher","first-page":"523","DOI":"10.3390\/drones8100523","volume":"8","author":"S Wang","year":"2024","unstructured":"Wang, S., Jiang, H., Yang, J., Ma, X., Chen, J.: Amfef-detr: An end-to-end adaptive multi-scale feature extraction and fusion object detection network based on uav aerial images. Drones 8(10), 523 (2024)","journal-title":"Drones"},{"issue":"4","key":"4924_CR17","doi-asserted-by":"publisher","first-page":"302","DOI":"10.1007\/s11760-025-03850-0","volume":"19","author":"J Li","year":"2025","unstructured":"Li, J., Zheng, C., Chen, P., Zhang, J., Wang, B.: Small object detection in uav imagery based on channel-spatial fusion cross attention. SIViP 19(4), 302 (2025). https:\/\/doi.org\/10.1007\/s11760-025-03850-0","journal-title":"SIViP"},{"issue":"17","key":"4924_CR18","doi-asserted-by":"publisher","first-page":"5496","DOI":"10.3390\/s24175496","volume":"24","author":"Y Kong","year":"2024","unstructured":"Kong, Y., Shang, X., Jia, S.: Drone-detr: Efficient small object detection for remote sensing image using enhanced rt-detr model. Sensors 24(17), 5496 (2024)","journal-title":"Sensors"},{"key":"4924_CR19","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2025.113539","volume":"318","author":"P Shi","year":"2025","unstructured":"Shi, P., Dong, X., Ge, R., Liu, Z., Yang, A.: Dp-m3d: Monocular 3d object detection algorithm with depth perception capability. Knowl.-Based Syst. 318, 113539 (2025)","journal-title":"Knowl.-Based Syst."},{"key":"4924_CR20","doi-asserted-by":"publisher","DOI":"10.1016\/j.displa.2024.102814","volume":"84","author":"X Dong","year":"2024","unstructured":"Dong, X., Shi, P., Qi, H., Yang, A., Liang, T.: Ts-bev: Bev object detection algorithm based on temporal-spatial feature fusion. Displays 84, 102814 (2024). https:\/\/doi.org\/10.1016\/j.displa.2024.102814","journal-title":"Displays"},{"issue":"1","key":"4924_CR21","doi-asserted-by":"publisher","first-page":"1947","DOI":"10.1177\/03611981241258753","volume":"2679","author":"X Dong","year":"2025","unstructured":"Dong, X., Shi, P., Liang, T., Yang, A.: Ctaffnet: Cnn-transformer adaptive feature fusion object detection algorithm for complex traffic scenarios. Transp. Res. Rec. 2679(1), 1947\u20131965 (2025)","journal-title":"Transp. Res. Rec."},{"key":"4924_CR22","doi-asserted-by":"crossref","unstructured":"Sunkara, R., Luo, T.: No more strided convolutions or pooling: A new cnn building block for low-resolution images and small objects. In: Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 443\u2013459 (2022). Springer","DOI":"10.1007\/978-3-031-26409-2_27"},{"key":"4924_CR23","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"4924_CR24","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y., Doll\u00e1r, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2117\u20132125 (2017)","DOI":"10.1109\/CVPR.2017.106"},{"key":"4924_CR25","doi-asserted-by":"crossref","unstructured":"Liu, S., Qi, L., Qin, H., Shi, J., Jia, J.: Path aggregation network for instance segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8759\u20138768 (2018)","DOI":"10.1109\/CVPR.2018.00913"},{"key":"4924_CR26","doi-asserted-by":"crossref","unstructured":"Tang, S., Zhang, S., Fang, Y.: Hic-yolov5: Improved yolov5 for small object detection. In: 2024 IEEE International Conference on Robotics and Automation (ICRA), pp. 6614\u20136619 (2024). IEEE","DOI":"10.1109\/ICRA57147.2024.10610273"},{"issue":"2","key":"4924_CR27","doi-asserted-by":"publisher","first-page":"143","DOI":"10.3390\/drones9020143","volume":"9","author":"Y Liu","year":"2025","unstructured":"Liu, Y., He, M., Hui, B.: Eso-detr: An improved real-time detection transformer model for enhanced small object detection in uav imagery. Drones 9(2), 143 (2025)","journal-title":"Drones"},{"key":"4924_CR28","doi-asserted-by":"crossref","unstructured":"Tan, M., Pang, R., Le, Q.V.: Efficientdet: Scalable and efficient object detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10781\u201310790 (2020)","DOI":"10.1109\/CVPR42600.2020.01079"},{"key":"4924_CR29","doi-asserted-by":"crossref","unstructured":"Ding, X., Zhang, X., Ma, N., Han, J., Ding, G., Sun, J.: Repvgg: Making vgg-style convnets great again. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13733\u201313742 (2021)","DOI":"10.1109\/CVPR46437.2021.01352"},{"key":"4924_CR30","unstructured":"Du, D., Zhu, P., Wen, L., Bian, X., Lin, H., Hu, Q., Peng, T., Zheng, J., Wang, X., Zhang, Y., et al. Visdrone-det2019: The vision meets drone object detection in image challenge results. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision Workshops, pp. 0\u20130 (2019)"},{"key":"4924_CR31","unstructured":"Powers, D.M.: Evaluation: from precision, recall and f-measure to roc, informedness, markedness and correlation. arXiv:2010.16061 (2020)"},{"key":"4924_CR32","unstructured":"Lei, M., Li, S., Wu, Y., Hu, H., Zhou, Y., Zheng, X., Ding, G., Du, S., Wu, Z., Gao, Y.: Yolov13: Real-time object detection with hypergraph-enhanced adaptive visual perception. arXiv:2506.17733 (2025)"},{"key":"4924_CR33","unstructured":"Lv, W., Zhao, Y., Chang, Q., Huang, K., Wang, G., Liu, Y.: Rt-detrv2: Improved baseline with bag-of-freebies for real-time detection transformer. arXiv:2407.17140 (2024)"},{"key":"4924_CR34","unstructured":"Tian, Y., Ye, Q., Doermann, D.: Yolov12: Attention-centric real-time object detectors. arXiv:2502.12524 (2025)"},{"issue":"1","key":"4924_CR35","doi-asserted-by":"publisher","first-page":"227","DOI":"10.1038\/s41597-023-02066-6","volume":"10","author":"J Suo","year":"2023","unstructured":"Suo, J., Wang, T., Zhang, X., Chen, H., Zhou, W., Shi, W.: Hit-uav: A high-altitude infrared thermal dataset for unmanned aerial vehicle-based object detection. Scientific Data 10(1), 227 (2023)","journal-title":"Scientific Data"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04924-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-025-04924-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04924-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,3]],"date-time":"2025-12-03T00:55:47Z","timestamp":1764723347000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-025-04924-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,10]]},"references-count":35,"journal-issue":{"issue":"16","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["4924"],"URL":"https:\/\/doi.org\/10.1007\/s11760-025-04924-9","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"type":"print","value":"1863-1703"},{"type":"electronic","value":"1863-1711"}],"subject":[],"published":{"date-parts":[[2025,11,10]]},"assertion":[{"value":"23 July 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 September 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 October 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 November 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of Interest"}},{"value":"The authors declare no competing interests.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"1348"}}