{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T00:02:57Z","timestamp":1773273777109,"version":"3.50.1"},"reference-count":37,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2024,12,19]],"date-time":"2024-12-19T00:00:00Z","timestamp":1734566400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2024,12,19]],"date-time":"2024-12-19T00:00:00Z","timestamp":1734566400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"name":"Excellent scientific and technological innovation team project of Jiangsu Provincial Colleges and Universities"},{"name":"Start-up funds for scientific research of Wuxi University","award":["2023R001"],"award-info":[{"award-number":["2023R001"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2025,1]]},"DOI":"10.1007\/s40747-024-01687-7","type":"journal-article","created":{"date-parts":[[2024,12,19]],"date-time":"2024-12-19T09:21:40Z","timestamp":1734600100000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["A UAV perspective based lightweight target detection and tracking algorithm for intelligent transportation"],"prefix":"10.1007","volume":"11","author":[{"given":"Quan","family":"Wang","sequence":"first","affiliation":[]},{"given":"Guangfei","family":"Ye","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9434-4979","authenticated-orcid":false,"given":"Qidong","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Songyang","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Fengqing","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,19]]},"reference":[{"key":"1687_CR1","doi-asserted-by":"publisher","unstructured":"Liu W et al (2016) SSD: single shot MultiBox detector. In: Leibe B, Matas J, Sebe N, Welling M (eds) Computer vision \u2013 ECCV 2016. ECCV 2016. Lecture notes in computer science, vol 9905. Springer, Cham. https:\/\/doi.org\/10.1007\/978-3-319-46448-0_2","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"1687_CR2","doi-asserted-by":"publisher","unstructured":"Jocher G et al (2020) ultralytics\/yolov5: v3.1 - Bug fixes and performance improvements. https:\/\/doi.org\/10.5281\/ZENODO.4154370","DOI":"10.5281\/ZENODO.4154370"},{"key":"1687_CR3","doi-asserted-by":"publisher","unstructured":"Reis D, Kupec J, Hong J, Daoudi A (2023) Real-time flying object detection with YOLOv8. https:\/\/doi.org\/10.48550\/arXiv.2305.09972","DOI":"10.48550\/arXiv.2305.09972"},{"key":"1687_CR4","doi-asserted-by":"publisher","unstructured":"Gao P, Ji C-L, Yu T, Yuan R-Y (2024) YOLO-TLA: an efficient and lightweight small object detection model based on YOLOv5. https:\/\/doi.org\/10.48550\/arXiv.2402.14309","DOI":"10.48550\/arXiv.2402.14309"},{"key":"1687_CR5","unstructured":"Hong S, Roh B, Kim K-H, Cheon Y, Park M (2016) PVANet: lightweight deep neural networks for real-time object detection. arXiv:1611.08588"},{"key":"1687_CR6","doi-asserted-by":"publisher","first-page":"12135","DOI":"10.1002\/int.23079","volume":"37","author":"Y Liu","year":"2022","unstructured":"Liu Y, Zhang C, Wu W et al (2022) MiniYOLO: a lightweight object detection algorithm that realizes the trade-off between model size and detection accuracy. Int J Intell Syst 37:12135\u201312151. https:\/\/doi.org\/10.1002\/int.23079","journal-title":"Int J Intell Syst"},{"key":"1687_CR7","doi-asserted-by":"publisher","first-page":"12082","DOI":"10.1038\/s41598-022-16208-0","volume":"12","author":"L Cheng","year":"2022","unstructured":"Cheng L, Ji Y, Li C et al (2022) Improved SSD network for fast concealed object detection and recognition in passive terahertz security images. Sci Rep 12:12082. https:\/\/doi.org\/10.1038\/s41598-022-16208-0","journal-title":"Sci Rep"},{"key":"1687_CR8","doi-asserted-by":"publisher","unstructured":"Tan M, Pang R, Le QV (2020) EfficientDet: scalable and efficient object detection. https:\/\/doi.org\/10.48550\/arXiv.1911.09070","DOI":"10.48550\/arXiv.1911.09070"},{"key":"1687_CR9","doi-asserted-by":"publisher","unstructured":"Han K, Wang Y, Tian Q, Guo J, Xu C, Xu C (2020) GhostNet: more features from cheap operations. https:\/\/doi.org\/10.48550\/arXiv.1911.11907","DOI":"10.48550\/arXiv.1911.11907"},{"key":"1687_CR10","doi-asserted-by":"crossref","unstructured":"Bewley A, Ge Z, Ott L et al (2016) Simple online and realtime tracking. In: 2016 IEEE international conference on image processing (ICIP), pp 3464\u20133468","DOI":"10.1109\/ICIP.2016.7533003"},{"key":"1687_CR11","doi-asserted-by":"crossref","unstructured":"Wojke N, Bewley A, Paulus D (2017) Simple online and realtime tracking with a deep association metric. arXiv:1703.07402","DOI":"10.1109\/ICIP.2017.8296962"},{"key":"1687_CR12","doi-asserted-by":"publisher","unstructured":"Zou Y, Yang X, Yu Z, Kumar BVKV, Kautz J (2020) Joint disentangling and adaptation for cross-domain person re-identification. https:\/\/doi.org\/10.48550\/arXiv.2007.10315","DOI":"10.48550\/arXiv.2007.10315"},{"key":"1687_CR13","doi-asserted-by":"publisher","unstructured":"Sun P et al (2021) TransTrack: multiple object tracking with transformer. https:\/\/doi.org\/10.48550\/arXiv.2012.15460","DOI":"10.48550\/arXiv.2012.15460"},{"key":"1687_CR14","doi-asserted-by":"publisher","first-page":"3069","DOI":"10.1007\/s11263-021-01513-4","volume":"129","author":"Y Zhang","year":"2021","unstructured":"Zhang Y, Wang C, Wang X et al (2021) FairMOT: on the fairness of detection and re-identification in multiple object tracking. Int J Comput Vis 129:3069\u20133087. https:\/\/doi.org\/10.1007\/s11263-021-01513-4","journal-title":"Int J Comput Vis"},{"key":"1687_CR15","doi-asserted-by":"publisher","unstructured":"Zhou X, Koltun V, Kr\u00e4henb\u00fchl P (2020) Tracking objects as points. https:\/\/doi.org\/10.48550\/arXiv.2004.01177","DOI":"10.48550\/arXiv.2004.01177"},{"key":"1687_CR16","doi-asserted-by":"publisher","first-page":"3782","DOI":"10.3390\/s23073782","volume":"23","author":"H Hou","year":"2023","unstructured":"Hou H, Shen C, Zhang X, Gao W (2023) CSMOT: make one-shot multi-object tracking in crowded scenes great again. Sensors 23:3782. https:\/\/doi.org\/10.3390\/s23073782","journal-title":"Sensors"},{"key":"1687_CR17","doi-asserted-by":"publisher","first-page":"3551","DOI":"10.3390\/rs15143551","volume":"15","author":"C Yu","year":"2023","unstructured":"Yu C, Feng Z, Wu Z et al (2023) HB-YOLO: an improved YOLOv7 algorithm for dim-object tracking in satellite remote sensing videos. Remote Sens 15:3551. https:\/\/doi.org\/10.3390\/rs15143551","journal-title":"Remote Sens"},{"key":"1687_CR18","doi-asserted-by":"publisher","unstructured":"Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: unified, real-time object detection. https:\/\/doi.org\/10.48550\/arXiv.1506.02640","DOI":"10.48550\/arXiv.1506.02640"},{"key":"1687_CR19","doi-asserted-by":"publisher","unstructured":"Bochkovskiy A, Wang C-Y, Liao H-YM (2020) YOLOv4: optimal speed and accuracy of object detection. https:\/\/doi.org\/10.48550\/arXiv.2004.10934","DOI":"10.48550\/arXiv.2004.10934"},{"key":"1687_CR20","doi-asserted-by":"publisher","first-page":"8574","DOI":"10.1109\/TCYB.2021.3095305","volume":"52","author":"Z Zheng","year":"2022","unstructured":"Zheng Z, Wang P, Ren D et al (2022) Enhancing geometric factors in model learning and inference for object detection and instance segmentation. IEEE Trans Cybern 52:8574\u20138586. https:\/\/doi.org\/10.1109\/TCYB.2021.3095305","journal-title":"IEEE Trans Cybern"},{"key":"1687_CR21","doi-asserted-by":"crossref","unstructured":"Feng C, Zhong Y, Gao Y et al (2021) TOOD: task-aligned one-stage object detection. In: 2021 IEEE\/CVF international conference on computer vision (ICCV), pp 3490\u20133499","DOI":"10.1109\/ICCV48922.2021.00349"},{"key":"1687_CR22","unstructured":"Welch G, Bishop G (2006) An introduction to the Kalman filter"},{"key":"1687_CR23","doi-asserted-by":"publisher","first-page":"2","DOI":"10.1002\/nav.20056","volume":"52","author":"A Frank","year":"2005","unstructured":"Frank A (2005) On Kuhn\u2019s Hungarian method\u2014a tribute from Hungary. Naval Res Logist (NRL) 52:2\u20135. https:\/\/doi.org\/10.1002\/nav.20056","journal-title":"Naval Res Logist (NRL)"},{"key":"1687_CR24","doi-asserted-by":"crossref","unstructured":"Li Y, Hou Q, Zheng Z, Cheng M-M, Yang J, Li X (2023) Large selective kernel network for remote sensing object detection. arXiv:2303.09030","DOI":"10.1109\/ICCV51070.2023.01540"},{"key":"1687_CR25","doi-asserted-by":"publisher","unstructured":"Wang C-Y, Liao H-YM, Yeh I-H, Wu Y-H, Chen P-Y, Hsieh J-W (2019) CSPNet: a new backbone that can enhance learning capability of CNN. https:\/\/doi.org\/10.48550\/arXiv.1911.11929","DOI":"10.48550\/arXiv.1911.11929"},{"key":"1687_CR26","doi-asserted-by":"crossref","unstructured":"Chen J, Kao S, He H et al (2023) Run, don\u2019t walk: chasing higher FLOPS for faster neural networks. In: 2023 IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp 12021\u201312031","DOI":"10.1109\/CVPR52729.2023.01157"},{"key":"1687_CR27","doi-asserted-by":"crossref","unstructured":"Neubeck A, Van Gool L (2006) Efficient non-maximum suppression. In: 18th international conference on pattern recognition (ICPR\u201906), pp 850\u2013855","DOI":"10.1109\/ICPR.2006.479"},{"key":"1687_CR28","doi-asserted-by":"publisher","unstructured":"Zhang Y et al (2022) ByteTrack: multi-object tracking by associating every detection box. https:\/\/doi.org\/10.48550\/arXiv.2110.06864","DOI":"10.48550\/arXiv.2110.06864"},{"key":"1687_CR29","doi-asserted-by":"publisher","unstructured":"Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Commun ACM 24(6). https:\/\/doi.org\/10.1145\/358669.358692. Accessed 13 Apr 2024","DOI":"10.1145\/358669.358692"},{"key":"1687_CR30","doi-asserted-by":"crossref","unstructured":"Du D, Zhu P, Wen L et al (2019) VisDrone-DET2019: the vision meets drone object detection in image challenge results. In: 2019 IEEE\/CVF international conference on computer vision workshop (ICCVW), pp 213\u2013226","DOI":"10.1109\/ICCVW.2019.00030"},{"key":"1687_CR31","doi-asserted-by":"crossref","unstructured":"Wen L, Zhu P, Du D et al (2019) VisDrone-MOT2019: the vision meets drone multiple object tracking challenge results. In: 2019 IEEE\/CVF international conference on computer vision workshop (ICCVW), pp 189\u2013198","DOI":"10.1109\/ICCVW.2019.00028"},{"key":"1687_CR32","doi-asserted-by":"crossref","unstructured":"Cao J, Pang J, Weng X, Khirodkar R, Kitani K (2023) Observation-centric SORT: rethinking SORT for robust multi-object tracking. arXiv:2203.14360","DOI":"10.1109\/CVPR52729.2023.00934"},{"key":"1687_CR33","doi-asserted-by":"crossref","unstructured":"Du Y et al (2024) StrongSORT: make DeepSORT great again. arXiv:2202.13514","DOI":"10.1109\/TMM.2023.3240881"},{"key":"1687_CR34","doi-asserted-by":"publisher","unstructured":"Xu Y, Osep A, Ban Y, Horaud R, Leal-Taixe L, Alameda-Pineda X (2019) How to train your deep multi-object tracker. https:\/\/doi.org\/10.48550\/arXiv.1906.06618","DOI":"10.48550\/arXiv.1906.06618"},{"key":"1687_CR35","doi-asserted-by":"crossref","unstructured":"Yang F, Odashima S, Masui S, Jiang S (2024) Hard to track objects with irregular motions and similar appearances? Make it easier by buffering the matching space. arXiv:2211.14317","DOI":"10.1109\/WACV56688.2023.00478"},{"key":"1687_CR36","doi-asserted-by":"crossref","unstructured":"Liu S, Li X, Lu H, He Y (2022) Multi-object tracking meets moving UAV. In: 2022 IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp 8866\u20138875","DOI":"10.1109\/CVPR52688.2022.00867"},{"key":"1687_CR37","doi-asserted-by":"publisher","unstructured":"Aharon N, Orfaig R, Bobrovsky B-Z (2022) BoT-SORT: robust associations multi-pedestrian tracking. https:\/\/doi.org\/10.48550\/arXiv.2206.14651","DOI":"10.48550\/arXiv.2206.14651"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01687-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-024-01687-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01687-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,30]],"date-time":"2025-01-30T20:17:09Z","timestamp":1738268229000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-024-01687-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,19]]},"references-count":37,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,1]]}},"alternative-id":["1687"],"URL":"https:\/\/doi.org\/10.1007\/s40747-024-01687-7","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,19]]},"assertion":[{"value":"23 April 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 November 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 December 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper. The authors declare the following financial interests\/personal relationships which may be considered as potential competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"73"}}