{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T03:29:43Z","timestamp":1773890983502,"version":"3.50.1"},"reference-count":75,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100017610","name":"Shenzhen Science and Technology Innovation Program","doi-asserted-by":"publisher","award":["JCYJ20240813104843058"],"award-info":[{"award-number":["JCYJ20240813104843058"]}],"id":[{"id":"10.13039\/501100017610","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2025YFE0101100"],"award-info":[{"award-number":["2025YFE0101100"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Engineering Applications of Artificial Intelligence"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1016\/j.engappai.2025.113463","type":"journal-article","created":{"date-parts":[[2025,12,5]],"date-time":"2025-12-05T17:30:51Z","timestamp":1764955851000},"page":"113463","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"PA","title":["A cascaded border-aware network for visual tracking"],"prefix":"10.1016","volume":"165","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-1610-5071","authenticated-orcid":false,"given":"Qun","family":"Li","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1648-0227","authenticated-orcid":false,"given":"Haijun","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Kai","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Zhili","family":"Zhou","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.engappai.2025.113463_b1","doi-asserted-by":"crossref","unstructured":"Bai, Y., Zhao, Z., Gong, Y., Wei, X., 2024. Artrackv2: Prompting autoregressive tracker where to look and how to describe. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 19048\u201319057.","DOI":"10.1109\/CVPR52733.2024.01802"},{"key":"10.1016\/j.engappai.2025.113463_b2","series-title":"Proceedings of the European Conference on Computer Vision","first-page":"205","article-title":"Know your surroundings: Exploiting scene information for object tracking","author":"Bhat","year":"2020"},{"key":"10.1016\/j.engappai.2025.113463_b3","doi-asserted-by":"crossref","unstructured":"Cai, Y., Liu, J., Tang, J., Wu, G., 2023. Robust object modeling for visual tracking. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 9589\u20139600.","DOI":"10.1109\/ICCV51070.2023.00879"},{"key":"10.1016\/j.engappai.2025.113463_b4","series-title":"Proceedings of the European Conference on Computer Vision","first-page":"213","article-title":"End-to-end object detection with transformers","author":"Carion","year":"2020"},{"key":"10.1016\/j.engappai.2025.113463_b5","first-page":"4838","article-title":"Improving visual object tracking through visual prompting","volume":"37","author":"Chen","year":"2025","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.engappai.2025.113463_b6","series-title":"Proceedings of the European Conference on Computer Vision","first-page":"375","article-title":"Backbone is all your need: A simplified architecture for visual object tracking","author":"Chen","year":"2022"},{"key":"10.1016\/j.engappai.2025.113463_b7","doi-asserted-by":"crossref","unstructured":"Chen, X., Peng, H., Wang, D., Lu, H., Hu, H., 2023. Seqtrack: Sequence to sequence learning for visual object tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 14572\u201314581.","DOI":"10.1109\/CVPR52729.2023.01400"},{"key":"10.1016\/j.engappai.2025.113463_b8","doi-asserted-by":"crossref","DOI":"10.1016\/j.cviu.2022.103547","article-title":"Fully convolutional online tracking","volume":"224","author":"Cui","year":"2022","journal-title":"Comput. Vis. Image Understanding"},{"key":"10.1016\/j.engappai.2025.113463_b9","doi-asserted-by":"crossref","unstructured":"Cui, Y., Jiang, C., Wang, L., Wu, G., 2022b. Mixformer: End-to-end tracking with iterative mixed attention. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 13608\u201313618.","DOI":"10.1109\/CVPR52688.2022.01324"},{"key":"10.1016\/j.engappai.2025.113463_b10","series-title":"Advances in Neural Information Processing Systems","article-title":"Mixformerv2: Efficient fully transformer tracking","volume":"vol. 36","author":"Cui","year":"2024"},{"key":"10.1016\/j.engappai.2025.113463_b11","doi-asserted-by":"crossref","unstructured":"Danelljan, M., Gool, L.V., Timofte, R., 2020. Probabilistic regression for visual tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 7183\u20137192.","DOI":"10.1109\/CVPR42600.2020.00721"},{"key":"10.1016\/j.engappai.2025.113463_b12","series-title":"An image is worth 16x16 words: Transformers for image recognition at scale","author":"Dosovitskiy","year":"2020"},{"key":"10.1016\/j.engappai.2025.113463_b13","doi-asserted-by":"crossref","unstructured":"Duan, K., Bai, S., Xie, L., Qi, H., Huang, Q., Tian, Q., 2019. Centernet: Keypoint triplets for object detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 6569\u20136578.","DOI":"10.1109\/ICCV.2019.00667"},{"key":"10.1016\/j.engappai.2025.113463_b14","doi-asserted-by":"crossref","unstructured":"Fan, H., Lin, L., Yang, F., Chu, P., Deng, G., Yu, S., Bai, H., Xu, Y., Liao, C., Ling, H., 2019. Lasot: A high-quality benchmark for large-scale single object tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 5374\u20135383.","DOI":"10.1109\/CVPR.2019.00552"},{"key":"10.1016\/j.engappai.2025.113463_b15","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"13774","article-title":"Stmtrack: Template-free visual tracking with space\u2013time memory networks","author":"Fu","year":"2021"},{"key":"10.1016\/j.engappai.2025.113463_b16","doi-asserted-by":"crossref","unstructured":"Galoogahi, H.Kiani., Fagg, A., Huang, C., Ramanan, D., Lucey, S., 2017. Need for speed: A benchmark for higher frame rate object tracking. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 1125\u20131134.","DOI":"10.1109\/ICCV.2017.128"},{"key":"10.1016\/j.engappai.2025.113463_b17","series-title":"Proceedings of the European Conference on Computer Vision","first-page":"146","article-title":"Aiatrack: Attention in attention for transformer visual tracking","author":"Gao","year":"2022"},{"key":"10.1016\/j.engappai.2025.113463_b18","doi-asserted-by":"crossref","unstructured":"Gao, S., Zhou, C., Zhang, J., 2023. Generalized relation modeling for transformer tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 18686\u201318695.","DOI":"10.1109\/CVPR52729.2023.01792"},{"key":"10.1016\/j.engappai.2025.113463_b19","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2024.108461","article-title":"Visual tracking with pyramidal feature fusion and transformer based model predictor","volume":"133","author":"Gong","year":"2024","journal-title":"Eng. Appl. Artif. Intell."},{"key":"10.1016\/j.engappai.2025.113463_b20","doi-asserted-by":"crossref","unstructured":"Gopal, G.Y., Amer, M.A., 2024. Separable self and mixed attention transformers for efficient object tracking. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision. pp. 6708\u20136717.","DOI":"10.1109\/WACV57701.2024.00657"},{"key":"10.1016\/j.engappai.2025.113463_b21","first-page":"1","article-title":"Temporal context and environment-aware correlation filter for uav object tracking","volume":"62","author":"He","year":"2024","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"10.1016\/j.engappai.2025.113463_b22","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J., 2016. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 770\u2013778.","DOI":"10.1109\/CVPR.2016.90"},{"key":"10.1016\/j.engappai.2025.113463_b23","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, C., Xie, S., Li, Z., Wang, Z., 2023. Target-aware tracking with long-term context attention. In: Proceedings of the AAAI Conference on Artificial Intelligence. pp. 773\u2013780.","DOI":"10.1609\/aaai.v37i1.25155"},{"issue":"8","key":"10.1016\/j.engappai.2025.113463_b24","doi-asserted-by":"crossref","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","article-title":"Long short-term memory","volume":"9","author":"Hochreiter","year":"1997","journal-title":"Neural Comput."},{"key":"10.1016\/j.engappai.2025.113463_b25","doi-asserted-by":"crossref","unstructured":"Hong, L., Yan, S., Zhang, R., Li, W., Zhou, X., Guo, P., Jiang, K., Chen, Y., Li, J., Chen, Z., et al., 2024. Onetracker: Unifying visual object tracking with foundation models and efficient tuning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 19079\u201319091.","DOI":"10.1109\/CVPR52733.2024.01805"},{"key":"10.1016\/j.engappai.2025.113463_b26","doi-asserted-by":"crossref","first-page":"1562","DOI":"10.1109\/TPAMI.2019.2957464","article-title":"Got-10k: A large high-diversity benchmark for generic object tracking in the wild","volume":"43","author":"Huang","year":"2019","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.engappai.2025.113463_b27","series-title":"European Conference on Computer Vision","first-page":"709","article-title":"Visual prompt tuning","author":"Jia","year":"2022"},{"key":"10.1016\/j.engappai.2025.113463_b28","doi-asserted-by":"crossref","unstructured":"Kang, B., Chen, X., Wang, D., Peng, H., Lu, H., 2023. Exploring lightweight hierarchical vision transformers for efficient visual tracking. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 9612\u20139621.","DOI":"10.1109\/ICCV51070.2023.00881"},{"key":"10.1016\/j.engappai.2025.113463_b29","series-title":"Advances in Neural Information Processing Systems","article-title":"Zoomtrack: target-aware non-uniform resizing for efficient visual tracking","volume":"vol. 36","author":"Kou","year":"2024"},{"key":"10.1016\/j.engappai.2025.113463_b30","doi-asserted-by":"crossref","unstructured":"Law, H., Deng, J., 2018. Cornernet: Detecting objects as paired keypoints. In: Proceedings of the European Conference on Computer Vision. pp. 734\u2013750.","DOI":"10.1007\/978-3-030-01264-9_45"},{"key":"10.1016\/j.engappai.2025.113463_b31","doi-asserted-by":"crossref","unstructured":"Li, X., Huang, Y., He, Z., Wang, Y., Lu, H., Yang, M.H., 2023. Citetracker: Correlating image and text for visual tracking. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 9974\u20139983.","DOI":"10.1109\/ICCV51070.2023.00915"},{"key":"10.1016\/j.engappai.2025.113463_b32","doi-asserted-by":"crossref","unstructured":"Li, F., Zhang, H., Liu, S., Guo, J., Ni, L.M., Zhang, L., 2022. Dn-detr: Accelerate detr training by introducing query denoising. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 13619\u201313627.","DOI":"10.1109\/CVPR52688.2022.01325"},{"key":"10.1016\/j.engappai.2025.113463_b33","article-title":"Corpn: Corner prediction network for visual tracking","author":"Li","year":"2025","journal-title":"IEEE Trans. Consum. Electron."},{"key":"10.1016\/j.engappai.2025.113463_b34","doi-asserted-by":"crossref","first-page":"2365","DOI":"10.1109\/TCE.2024.3374239","article-title":"Ban-ima: A box adaptive network with iterative mixed attention for visual tracking","volume":"70","author":"Li","year":"2024","journal-title":"IEEE Trans. Consum. Electron."},{"key":"10.1016\/j.engappai.2025.113463_b35","series-title":"Advances in Neural Information Processing Systems","first-page":"16743","article-title":"Swintrack: A simple and strong baseline for transformer tracking","volume":"vol. 35","author":"Lin","year":"2022"},{"key":"10.1016\/j.engappai.2025.113463_b36","series-title":"Proceedings of the European Conference on Computer Vision","first-page":"740","article-title":"Microsoft coco: Common objects in context","author":"Lin","year":"2014"},{"key":"10.1016\/j.engappai.2025.113463_b37","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., Guo, B., 2021. Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 10012\u201310022.","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"10.1016\/j.engappai.2025.113463_b38","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2024.112736","article-title":"A transformer based visual tracker with restricted token interaction and knowledge distillation","volume":"307","author":"Liu","year":"2025","journal-title":"Knowl.-Based Syst."},{"key":"10.1016\/j.engappai.2025.113463_b39","doi-asserted-by":"crossref","unstructured":"Mayer, C., Danelljan, M., Bhat, G., Paul, M., Paudel, D.P., Yu, F., Gool, L.Van. and, 2022. Transforming model prediction for tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 8731\u20138740.","DOI":"10.1109\/CVPR52688.2022.00853"},{"key":"10.1016\/j.engappai.2025.113463_b40","doi-asserted-by":"crossref","unstructured":"Mayer, C., Danelljan, M., Paudel, D.P., Van Gool, L., 2021. Learning target candidate association to keep track of what not to track. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 13444\u201313454.","DOI":"10.1109\/ICCV48922.2021.01319"},{"key":"10.1016\/j.engappai.2025.113463_b41","series-title":"Proceedings of the European Conference on Computer Vision","first-page":"445","article-title":"A benchmark and simulator for uav tracking","author":"Mueller","year":"2016"},{"key":"10.1016\/j.engappai.2025.113463_b42","doi-asserted-by":"crossref","unstructured":"Muller, M., Bibi, A., Giancola, S., Alsubaihi, S., Ghanem, B., 2018. Trackingnet: A large-scale dataset and benchmark for object tracking in the wild. In: Proceedings of the European Conference on Computer Vision. pp. 300\u2013317.","DOI":"10.1007\/978-3-030-01246-5_19"},{"key":"10.1016\/j.engappai.2025.113463_b43","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2025.110482","article-title":"Visual object tracking using learnable target-aware token emphasis","volume":"149","author":"Park","year":"2025","journal-title":"Eng. Appl. Artif. Intell."},{"key":"10.1016\/j.engappai.2025.113463_b44","series-title":"Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August (2020) 23\u201328, Proceedings, Part I 16","first-page":"549","article-title":"Borderdet: Border feature for dense object detection","author":"Qiu","year":"2020"},{"key":"10.1016\/j.engappai.2025.113463_b45","doi-asserted-by":"crossref","unstructured":"Rezatofighi, H., Tsoi, N., Gwak, J., Sadeghian, A., Reid, I., Savarese, S., 2019. Generalized intersection over union: A metric and a loss for bounding box regression. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 658\u2013666.","DOI":"10.1109\/CVPR.2019.00075"},{"key":"10.1016\/j.engappai.2025.113463_b46","doi-asserted-by":"crossref","unstructured":"Song, Z., Yu, J., Chen, Y.P., Yang, W., 2022. Transformer tracking with cyclic shifting window attention. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 8791\u20138800.","DOI":"10.1109\/CVPR52688.2022.00859"},{"key":"10.1016\/j.engappai.2025.113463_b47","doi-asserted-by":"crossref","unstructured":"Tian, Z., Shen, C., Chen, H., He, T., 2019. FCOS: Fully convolutional one-stage object detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 9626\u20139635.","DOI":"10.1109\/ICCV.2019.00972"},{"key":"10.1016\/j.engappai.2025.113463_b48","series-title":"Advances in Neural Information Processing Systems","article-title":"Attention is all you need","volume":"vol. 30","author":"Vaswani","year":"2017"},{"key":"10.1016\/j.engappai.2025.113463_b49","doi-asserted-by":"crossref","unstructured":"Voigtlaender, P., Luiten, J., Torr, P.H., Leibe, B., 2020. Siam r-cnn: Visual tracking by re-detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 6578\u20136588.","DOI":"10.1109\/CVPR42600.2020.00661"},{"key":"10.1016\/j.engappai.2025.113463_b50","doi-asserted-by":"crossref","first-page":"899","DOI":"10.1007\/s11263-022-01732-3","article-title":"Watb: wild animal tracking benchmark","volume":"131","author":"Wang","year":"2023","journal-title":"Int. J. Comput. Vis."},{"key":"10.1016\/j.engappai.2025.113463_b51","doi-asserted-by":"crossref","first-page":"3159","DOI":"10.1109\/TPAMI.2025.3529926","article-title":"Omnitracker: Unifying visual object tracking by tracking-with-detection","volume":"47","author":"Wang","year":"2025","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.engappai.2025.113463_b52","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2024.108329","article-title":"Dynamic region-aware transformer backbone network for visual tracking","volume":"133","author":"Wang","year":"2024","journal-title":"Eng. Appl. Artif. Intell."},{"key":"10.1016\/j.engappai.2025.113463_b53","doi-asserted-by":"crossref","first-page":"326","DOI":"10.1109\/TMM.2023.3264851","article-title":"Cmat: Integrating convolution mixer and self-attention for visual tracking","volume":"26","author":"Wang","year":"2023","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.engappai.2025.113463_b54","doi-asserted-by":"crossref","unstructured":"Wang, N., Zhou, W., Wang, J., Li, H., 2021. Transformer meets tracker: Exploiting temporal context for robust visual tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 1571\u20131580.","DOI":"10.1109\/CVPR46437.2021.00162"},{"key":"10.1016\/j.engappai.2025.113463_b55","doi-asserted-by":"crossref","unstructured":"Wei, X., Bai, Y., Zheng, Y., Shi, D., Gong, Y., 2023. Autoregressive visual tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 9697\u20139706.","DOI":"10.1109\/CVPR52729.2023.00935"},{"key":"10.1016\/j.engappai.2025.113463_b56","doi-asserted-by":"crossref","unstructured":"Wu, Y., Lim, J., Yang, M.H., 2013. Online object tracking: A benchmark. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 2411\u20132418.","DOI":"10.1109\/CVPR.2013.312"},{"key":"10.1016\/j.engappai.2025.113463_b57","doi-asserted-by":"crossref","first-page":"1834","DOI":"10.1109\/TPAMI.2014.2388226","article-title":"Object tracking benchmark","volume":"37","author":"Wu","year":"2015","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.engappai.2025.113463_b58","doi-asserted-by":"crossref","unstructured":"Xie, F., Wang, C., Wang, G., Cao, Y., Yang, W., Zeng, W., 2022. Correlation-aware deep tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 8751\u20138760.","DOI":"10.1109\/CVPR52688.2022.00855"},{"key":"10.1016\/j.engappai.2025.113463_b59","doi-asserted-by":"crossref","unstructured":"Xie, J., Zhong, B., Liang, Q., Li, N., Mo, Z., Song, S., 2025. Robust tracking via mamba-based context-aware token learning. In: Proceedings of the AAAI Conference on Artificial Intelligence. pp. 8727\u20138735.","DOI":"10.1609\/aaai.v39i8.32943"},{"key":"10.1016\/j.engappai.2025.113463_b60","doi-asserted-by":"crossref","unstructured":"Yan, B., Jiang, Y., Wu, J., Wang, D., Luo, P., Yuan, Z., Lu, H., 2023. Universal instance perception as object discovery and retrieval. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 15325\u201315336.","DOI":"10.1109\/CVPR52729.2023.01471"},{"key":"10.1016\/j.engappai.2025.113463_b61","series-title":"Proceedings of the European Conference on Computer Vision","first-page":"152","article-title":"Learning dynamic memory networks for object tracking","author":"Yang","year":"2018"},{"key":"10.1016\/j.engappai.2025.113463_b62","doi-asserted-by":"crossref","DOI":"10.1016\/j.neunet.2024.106380","article-title":"Defort: Deformable transformer for visual tracking","volume":"176","author":"Yang","year":"2024","journal-title":"Neural Netw."},{"key":"10.1016\/j.engappai.2025.113463_b63","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2024.111278","article-title":"Adaptively bypassing vision transformer blocks for efficient visual tracking","volume":"161","author":"Yang","year":"2025","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.engappai.2025.113463_b64","doi-asserted-by":"crossref","first-page":"377","DOI":"10.1109\/TCE.2023.3251407","article-title":"Bandt: A border-aware network with deformable transformers for visual tracking","volume":"69","author":"Yang","year":"2023","journal-title":"IEEE Trans. Consum. Electron."},{"key":"10.1016\/j.engappai.2025.113463_b65","series-title":"Proceedings of the European Conference on Computer Vision","first-page":"341","article-title":"Joint feature learning and relation modeling for tracking: A one-stream framework","author":"Ye","year":"2022"},{"key":"10.1016\/j.engappai.2025.113463_b66","doi-asserted-by":"crossref","unstructured":"Ye, M., Ke, L., Li, S., Tai, Y.W., Tang, C.K., Danelljan, M., Yu, F., 2023. Cascade-detr: delving into high-quality universal object detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 6704\u20136714.","DOI":"10.1109\/ICCV51070.2023.00617"},{"key":"10.1016\/j.engappai.2025.113463_b67","doi-asserted-by":"crossref","unstructured":"Yu, J., Jiang, Y., Wang, Z., Cao, Z., Huang, T., 2016. Unitbox: An advanced object detection network. In: Proceedings of the 24th ACM International Conference on Multimedia. pp. 516\u2013520.","DOI":"10.1145\/2964284.2967274"},{"key":"10.1016\/j.engappai.2025.113463_b68","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Peng, H., Fu, J., Li, B., Hu, W., 2020. Ocean: Object-aware anchor-free tracking. In: Proceedings of the European Conference on Computer Vision, Springer. pp. 771\u2013787.","DOI":"10.1007\/978-3-030-58589-1_46"},{"key":"10.1016\/j.engappai.2025.113463_b69","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2025.110511","article-title":"Learning adaptive distractor-aware-suppression appearance model for visual tracking","volume":"150","author":"Zhang","year":"2025","journal-title":"Eng. Appl. Artif. Intell."},{"key":"10.1016\/j.engappai.2025.113463_b70","doi-asserted-by":"crossref","unstructured":"Zhao, J., Edstedt, J., Felsberg, M., Wang, D., Lu, H., 2024. Leveraging the power of data augmentation for transformer-based tracking. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision. pp. 6469\u20136478.","DOI":"10.1109\/WACV57701.2024.00634"},{"key":"10.1016\/j.engappai.2025.113463_b71","series-title":"Proceedings of the European Conference on Computer Vision","first-page":"759","article-title":"Learning feature embeddings for discriminant model based tracking","author":"Zheng","year":"2020"},{"key":"10.1016\/j.engappai.2025.113463_b72","doi-asserted-by":"crossref","first-page":"2908","DOI":"10.1109\/TCSVT.2021.3103063","article-title":"Target-aware state estimation for visual tracking","volume":"32","author":"Zhou","year":"2021","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.engappai.2025.113463_b73","doi-asserted-by":"crossref","unstructured":"Zhou, L., Zhou, Z., Mao, K., He, Z., 2023. Joint visual grounding and tracking with natural language specification. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 23151\u201323160.","DOI":"10.1109\/CVPR52729.2023.02217"},{"key":"10.1016\/j.engappai.2025.113463_b74","series-title":"Deformable detr: Deformable transformers for end-to-end object detection","author":"Zhu","year":"2020"},{"key":"10.1016\/j.engappai.2025.113463_b75","doi-asserted-by":"crossref","unstructured":"Zhu, J., Tang, H., Chen, X., Wang, X., Wang, D., Lu, H., 2025. Two-stream beats one-stream: asymmetric siamese network for efficient visual tracking. In: Proceedings of the AAAI Conference on Artificial Intelligence. pp. 10959\u201310967.","DOI":"10.1609\/aaai.v39i10.33191"}],"container-title":["Engineering Applications of Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0952197625034943?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0952197625034943?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T00:28:45Z","timestamp":1773880125000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0952197625034943"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2]]},"references-count":75,"alternative-id":["S0952197625034943"],"URL":"https:\/\/doi.org\/10.1016\/j.engappai.2025.113463","relation":{},"ISSN":["0952-1976"],"issn-type":[{"value":"0952-1976","type":"print"}],"subject":[],"published":{"date-parts":[[2026,2]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"A cascaded border-aware network for visual tracking","name":"articletitle","label":"Article Title"},{"value":"Engineering Applications of Artificial Intelligence","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.engappai.2025.113463","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"113463"}}