{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T22:11:47Z","timestamp":1776204707977,"version":"3.50.1"},"reference-count":39,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100003787","name":"Natural Science Foundation of Hebei Province","doi-asserted-by":"publisher","award":["F2025203059"],"award-info":[{"award-number":["F2025203059"]}],"id":[{"id":"10.13039\/501100003787","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003787","name":"Natural Science Foundation of Hebei Province","doi-asserted-by":"publisher","award":["F2024203112"],"award-info":[{"award-number":["F2024203112"]}],"id":[{"id":"10.13039\/501100003787","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003787","name":"Natural Science Foundation of Hebei Province","doi-asserted-by":"publisher","award":["F2025203057"],"award-info":[{"award-number":["F2025203057"]}],"id":[{"id":"10.13039\/501100003787","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012165","name":"Key Technologies Research and Development Program","doi-asserted-by":"publisher","award":["2023YFB2504400"],"award-info":[{"award-number":["2023YFB2504400"]}],"id":[{"id":"10.13039\/501100012165","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003482","name":"Hebei Province Department of Education","doi-asserted-by":"publisher","award":["BJ2026018"],"award-info":[{"award-number":["BJ2026018"]}],"id":[{"id":"10.13039\/501100003482","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003482","name":"Hebei Province Department of Education","doi-asserted-by":"publisher","award":["A2025010"],"award-info":[{"award-number":["A2025010"]}],"id":[{"id":"10.13039\/501100003482","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52472440"],"award-info":[{"award-number":["52472440"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U24A6008"],"award-info":[{"award-number":["U24A6008"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Internet of Things"],"published-print":{"date-parts":[[2026,5]]},"DOI":"10.1016\/j.iot.2026.101942","type":"journal-article","created":{"date-parts":[[2026,4,4]],"date-time":"2026-04-04T15:23:07Z","timestamp":1775316187000},"page":"101942","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["A Dual-Adaptive framework for 3D multi-Object tracking with fast hierarchical association in IoT intelligent vehicles"],"prefix":"10.1016","volume":"37","author":[{"ORCID":"https:\/\/orcid.org\/0009-0000-4303-441X","authenticated-orcid":false,"given":"Hongyu","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Baicang","family":"Guo","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3086-1333","authenticated-orcid":false,"given":"Lisheng","family":"Jin","sequence":"additional","affiliation":[]},{"given":"Chenyi","family":"Yue","sequence":"additional","affiliation":[]},{"given":"Xingchen","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Yewei","family":"Shi","sequence":"additional","affiliation":[]},{"given":"Tao","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Yang","family":"He","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.iot.2026.101942_bib0001","doi-asserted-by":"crossref","DOI":"10.1016\/j.neucom.2024.127635","article-title":"A review of object tracking methods: from general field to autonomous vehicles","volume":"585","author":"Cao","year":"2024","journal-title":"Neurocomputing"},{"key":"10.1016\/j.iot.2026.101942_bib0002","doi-asserted-by":"crossref","DOI":"10.1016\/j.measurement.2025.117864","article-title":"A dynamic-confidence 3D multi-object tracking method based on spatio-temporal association","volume":"256","author":"Zeng","year":"2025","journal-title":"Measurement"},{"key":"10.1016\/j.iot.2026.101942_bib0003","doi-asserted-by":"crossref","DOI":"10.1016\/j.sigpro.2025.109958","article-title":"Escapetrack: multi-object tracking with estimated camera parameters","volume":"234","author":"Yi","year":"2025","journal-title":"Signal Process."},{"key":"10.1016\/j.iot.2026.101942_bib0004","article-title":"Surface multiple object tracking: an accurate HAT-YOLOv8-ADT tracking model","author":"Lin","year":"2025","journal-title":"IEEE Internet Things J."},{"issue":"2","key":"10.1016\/j.iot.2026.101942_bib0005","doi-asserted-by":"crossref","first-page":"103","DOI":"10.1109\/MVT.2019.2892497","article-title":"The impact of adverse weather conditions on autonomous vehicles: how rain, snow, fog, and hail affect the performance of a self-driving car","volume":"14","author":"Zang","year":"2019","journal-title":"IEEE Veh. Technol. Mag."},{"key":"10.1016\/j.iot.2026.101942_bib0006","first-page":"1","article-title":"Pillar3D-Former: a pillar-Based 3-D object detection and tracking method for autonomous driving scenes","volume":"73","author":"Tao","year":"2024","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.iot.2026.101942_bib0007","series-title":"RSJ International Conference on Intelligent Robots and Systems (IROS)","first-page":"9391","article-title":"Poly-mot: a polyhedral framework for 3d multi-object tracking. in 2023\u202fIEEE","author":"Li","year":"2023"},{"issue":"6","key":"10.1016\/j.iot.2026.101942_bib0008","doi-asserted-by":"crossref","first-page":"5668","DOI":"10.1109\/TITS.2021.3055616","article-title":"3D Multi-Object tracking in point clouds based on prediction confidence-Guided data association","volume":"23","author":"Wu","year":"2022","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"10.1016\/j.iot.2026.101942_bib0009","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.neucom.2021.05.011","article-title":"Recent advances of single-object tracking methods: a brief survey","volume":"455","author":"Zhang","year":"2021","journal-title":"Neurocomputing"},{"key":"10.1016\/j.iot.2026.101942_bib0010","series-title":"2020Del InsThinspace IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS)","first-page":"10359","article-title":"3D Multi-Object tracking: a baseline and new evaluation metrics","author":"Weng","year":"2020"},{"key":"10.1016\/j.iot.2026.101942_bib0011","series-title":"2021Del InsThinspace IEEE International Conference on Robotics and Automation (ICRA)","first-page":"14227","article-title":"Probabilistic 3D multi-Modal, multi-Object tracking for autonomous driving","author":"Chiu","year":"2021"},{"key":"10.1016\/j.iot.2026.101942_bib0012","series-title":"2021Del InsThinspace IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"11779","article-title":"Center-based 3D object detection and tracking","author":"Yin","year":"2021"},{"issue":"1","key":"10.1016\/j.iot.2026.101942_bib0013","doi-asserted-by":"crossref","first-page":"512","DOI":"10.1109\/TIV.2022.3158419","article-title":"3D Multi-Object tracking with adaptive cubature kalman filter for autonomous driving","volume":"8","author":"Guo","year":"2023","journal-title":"IEEE Trans. Intell. Veh."},{"key":"10.1016\/j.iot.2026.101942_bib0014","series-title":"European Conference on Computer Vision","first-page":"680","article-title":"Simpletrack: understanding and rethinking 3d multi-object tracking","author":"Pang","year":"2022"},{"issue":"5","key":"10.1016\/j.iot.2026.101942_bib0015","doi-asserted-by":"crossref","first-page":"4273","DOI":"10.1109\/LRA.2023.3323124","article-title":"ShaSTA: modeling shape and spatio-Temporal affinities for 3D multi-Object tracking","volume":"9","author":"Sadjadpour","year":"2024","journal-title":"IEEE Rob. Autom. Lett."},{"issue":"2","key":"10.1016\/j.iot.2026.101942_bib0016","doi-asserted-by":"crossref","first-page":"5103","DOI":"10.1109\/LRA.2022.3145952","article-title":"Learnable online graph representations for 3D multi-Object tracking","volume":"7","author":"Zaech","year":"2022","journal-title":"IEEE Rob. Autom. Lett."},{"key":"10.1016\/j.iot.2026.101942_bib0017","series-title":"European Conference on Computer Vision","first-page":"41","article-title":"PolarMOT: how far can geometric relations take us in 3D multi-object tracking?","author":"Kim","year":"2022"},{"key":"10.1016\/j.iot.2026.101942_bib0018","doi-asserted-by":"crossref","first-page":"15","DOI":"10.1109\/TSP.2023.3314275","article-title":"Neural enhanced belief propagation for multiobject tracking","volume":"72","author":"Liang","year":"2024","journal-title":"IEEE Trans. Signal Process."},{"key":"10.1016\/j.iot.2026.101942_bib0019","series-title":"2022Del InsThinspace IEEE Intelligent Vehicles Symposium (IV)","first-page":"852","article-title":"Transformers for multi-object tracking on point clouds","author":"Ruppel","year":"2022"},{"key":"10.1016\/j.iot.2026.101942_bib0020","series-title":"2020Del InsThinspace IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"6498","article-title":"GNN3Dmot: graph neural network for 3D multi-Object tracking with 2D-3D multi-Feature learning","author":"Weng","year":"2020"},{"key":"10.1016\/j.iot.2026.101942_bib0021","series-title":"2021Del InsThinspace IEEE International Conference on Robotics and Automation (ICRA)","first-page":"11315","article-title":"EagerMOT: 3D multi-Object tracking via sensor fusion","author":"Kim","year":"2021"},{"issue":"11","key":"10.1016\/j.iot.2026.101942_bib0022","doi-asserted-by":"crossref","first-page":"11981","DOI":"10.1109\/TITS.2023.3285651","article-title":"CAMO-MOT: Combined appearance-Motion optimization for 3D multi-Object tracking with camera-Lidar fusion","volume":"24","author":"Wang","year":"2023","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"10.1016\/j.iot.2026.101942_bib0023","doi-asserted-by":"crossref","DOI":"10.1016\/j.neucom.2025.131687","article-title":"Multi-modal 3D multi-object tracking with robust association and track drift compensation","author":"Xie","year":"2025","journal-title":"Neurocomputing"},{"issue":"12","key":"10.1016\/j.iot.2026.101942_bib0024","doi-asserted-by":"crossref","first-page":"21290","DOI":"10.1109\/JIOT.2025.3546639","article-title":"AHMOT: Adaptive kalman filtering and hierarchical data association for 3-D multiobject tracking in IoT-Enabled autonomous vehicles","volume":"12","author":"Jiang","year":"2025","journal-title":"IEEE Internet Things J."},{"issue":"2","key":"10.1016\/j.iot.2026.101942_bib0025","doi-asserted-by":"crossref","first-page":"1992","DOI":"10.1109\/TPAMI.2022.3168781","article-title":"Monocular quasi-Dense 3D object tracking","volume":"45","author":"Hu","year":"2023","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.iot.2026.101942_bib0026","series-title":"2018Del InsThinspace IEEE Winter Conference on Applications of Computer Vision (WACV)","first-page":"466","article-title":"Recurrent autoregressive networks for online multi-object tracking","author":"Fang","year":"2018"},{"key":"10.1016\/j.iot.2026.101942_bib0027","series-title":"2012Del InsThinspace IEEE Conference on Computer Vision and Pattern Recognition","first-page":"3354","article-title":"Are we ready for autonomous driving? the KITTI vision benchmark suite","author":"Geiger","year":"2012"},{"key":"10.1016\/j.iot.2026.101942_bib0028","series-title":"2020Del InsThinspace IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"2443","article-title":"Scalability in perception for autonomous driving: waymo open dataset","author":"Sun","year":"2020"},{"issue":"1","key":"10.1016\/j.iot.2026.101942_bib0029","article-title":"Evaluating multiple object tracking performance: the clear mot metrics","volume":"2008","author":"Bernardin","year":"2008","journal-title":"EURASIP J. Image Video Process."},{"key":"10.1016\/j.iot.2026.101942_bib0030","first-page":"1","article-title":"Casa: a cascade attention network for 3-D object detection from liDAR point clouds","volume":"60","author":"Wu","year":"2022","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"issue":"4","key":"10.1016\/j.iot.2026.101942_bib0031","doi-asserted-by":"crossref","first-page":"9707","DOI":"10.1109\/LRA.2022.3191558","article-title":"3D Multi-Object tracking using graph neural networks with cross-Edge modality attention","volume":"7","author":"B\u00fcchner","year":"2022","journal-title":"IEEE Rob. Autom. Lett."},{"key":"10.1016\/j.iot.2026.101942_bib0032","doi-asserted-by":"crossref","DOI":"10.1016\/j.sigpro.2024.109544","article-title":"3D Multi-object tracking based on informatic divergence-guided data association","volume":"222","author":"He","year":"2024","journal-title":"Signal Process."},{"issue":"3","key":"10.1016\/j.iot.2026.101942_bib0033","doi-asserted-by":"crossref","first-page":"8260","DOI":"10.1109\/LRA.2022.3187264","article-title":"DeepfusionMOT: a 3D multi-Object tracking framework based on camera-LiDAR fusion with deep association","volume":"7","author":"Wang","year":"2022","journal-title":"IEEE Rob. Autom. Lett."},{"issue":"5","key":"10.1016\/j.iot.2026.101942_bib0034","doi-asserted-by":"crossref","first-page":"4377","DOI":"10.1109\/LRA.2024.3379865","article-title":"PNAS-MOT: Multi-Modal object tracking with pareto neural architecture search","volume":"9","author":"Peng","year":"2024","journal-title":"IEEE Rob. Autom. Lett."},{"issue":"1","key":"10.1016\/j.iot.2026.101942_bib0035","doi-asserted-by":"crossref","first-page":"532","DOI":"10.1109\/LRA.2024.3511438","article-title":"A multi-Modal fusion-Based 3D multi-Object tracking framework with joint detection","volume":"10","author":"Wang","year":"2025","journal-title":"IEEE Rob. Autom. Lett."},{"key":"10.1016\/j.iot.2026.101942_bib0036","doi-asserted-by":"crossref","unstructured":"Q. Wang, Y. Chen, Z. Pang, N. Wang, Z. Zhang, Immortal tracker: tracklet never dies, arXiv preprint arXiv: 2111.13672(2021).","DOI":"10.31219\/osf.io\/nw3fy"},{"key":"10.1016\/j.iot.2026.101942_bib0037","series-title":"2023Del InsThinspace IEEE\/CVF International Conference on Computer Vision (ICCV)","first-page":"18481","article-title":"Trajectoryformer: 3D object tracking transformer with predictive trajectory hypotheses","author":"Chen","year":"2023"},{"key":"10.1016\/j.iot.2026.101942_bib0038","doi-asserted-by":"crossref","unstructured":"X. Li, D. Liu, Y. Wu, X. Wu, L. Zhao, J. Gao, Fast-poly: a fast polyhedral framework for 3d multi-object tracking, arXiv preprint arXiv: 2403.13443(2024).","DOI":"10.1109\/LRA.2024.3475882"},{"issue":"6","key":"10.1016\/j.iot.2026.101942_bib0039","first-page":"1015","article-title":"Semi solid-State liDAR object detection algorithm enhanced by feature stability enhancement","volume":"46","author":"Jin","year":"2024","journal-title":"Qiche Gongcheng\/Automot. Eng."}],"container-title":["Internet of Things"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S2542660526000727?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S2542660526000727?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T21:38:47Z","timestamp":1776202727000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S2542660526000727"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,5]]},"references-count":39,"alternative-id":["S2542660526000727"],"URL":"https:\/\/doi.org\/10.1016\/j.iot.2026.101942","relation":{},"ISSN":["2542-6605"],"issn-type":[{"value":"2542-6605","type":"print"}],"subject":[],"published":{"date-parts":[[2026,5]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"A Dual-Adaptive framework for 3D multi-Object tracking with fast hierarchical association in IoT intelligent vehicles","name":"articletitle","label":"Article Title"},{"value":"Internet of Things","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.iot.2026.101942","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"101942"}}