{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T15:33:30Z","timestamp":1772120010419,"version":"3.50.1"},"reference-count":43,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"the National Natural Science Foundation of China under Grant","award":["62273243"],"award-info":[{"award-number":["62273243"]}]},{"name":"Joint Fund Key Project of Science and Technology & Plan of Henan Province","award":["235200810022"],"award-info":[{"award-number":["235200810022"]}]},{"name":"the Distinguished Youth Science Foundation of Henan province of China under Grant","award":["242300421055"],"award-info":[{"award-number":["242300421055"]}]},{"name":"Henan Province Key & Project","award":["241111210400"],"award-info":[{"award-number":["241111210400"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2026,1]]},"DOI":"10.1007\/s11760-025-05083-7","type":"journal-article","created":{"date-parts":[[2026,1,16]],"date-time":"2026-01-16T18:30:15Z","timestamp":1768588215000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Target-background interaction prompt framework for target tracking"],"prefix":"10.1007","volume":"20","author":[{"given":"Huanlong","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Junlong","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Weiqiang","family":"Fu","sequence":"additional","affiliation":[]},{"given":"Linwei","family":"Li","sequence":"additional","affiliation":[]},{"given":"Wanguo","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yanfeng","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,16]]},"reference":[{"key":"5083_CR1","doi-asserted-by":"crossref","unstructured":"Dan, L., Shizhan, Y.: Application of visual tracking based on infrared target imaging in aerobics action evaluation system. Optical and Quantum Electronics 56(3) (2023)","DOI":"10.1007\/s11082-023-06016-9"},{"key":"5083_CR2","doi-asserted-by":"crossref","unstructured":"Peng, J., Cui, Y., Zhong, Z., An, Y.: Ore rock fragmentation calculation based on multi-modal fusion of point clouds and images. Appl. Sci. 13(23) (2023)","DOI":"10.3390\/app132312558"},{"key":"5083_CR3","doi-asserted-by":"crossref","unstructured":"Hu, K., Yang, W., Huang, W., Zhou, X., Cao, M., Ren, J., Tan, H.: Sequential fusion based multi-granularity consistency for space-time transformer tracking. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, pp. 12519\u201312527 (2024)","DOI":"10.1609\/aaai.v38i11.29145"},{"key":"5083_CR4","doi-asserted-by":"crossref","unstructured":"Chen, X., Peng, H., Wang, D., Lu, H., Hu, H.: Seqtrack: Sequence to sequence learning for visual object tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14572\u201314581 (2023)","DOI":"10.1109\/CVPR52729.2023.01400"},{"key":"5083_CR5","doi-asserted-by":"crossref","unstructured":"Mayer, C., Danelljan, M., Bhat, G., Paul, M., Paudel, D.P., Yu, F., Van\u00a0Gool, L.: Transforming model prediction for tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8731\u20138740 (2022)","DOI":"10.1109\/CVPR52688.2022.00853"},{"key":"5083_CR6","doi-asserted-by":"publisher","first-page":"11362","DOI":"10.1109\/TCSVT.2025.3578479","volume":"35","author":"J Wang","year":"2025","unstructured":"Wang, J., Chai, B., Zhou, L., Wang, Y.: Robust object tracking via long-range spatial representation and local feature enhancement. IEEE Trans. Circuits Syst. Video Technol. 35, 11362\u201311376 (2025)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"5083_CR7","doi-asserted-by":"crossref","unstructured":"Xie, J., Zhong, B., Liang, Q., Li, N., Mo, Z., Song, S.: Robust tracking via mamba-based context-aware token learning (2024)","DOI":"10.1609\/aaai.v39i8.32943"},{"key":"5083_CR8","doi-asserted-by":"crossref","unstructured":"Zheng, Y., Zhong, B., Liang, Q., Zhang, S., Li, G., Li, X., Ji, R.: Towards universal modal tracking with online dense temporal token learning. IEEE Transactions on Pattern Analysis and Machine Intelligence (2025)","DOI":"10.1109\/TPAMI.2025.3593543"},{"key":"5083_CR9","doi-asserted-by":"publisher","first-page":"131993","DOI":"10.1016\/j.optcom.2025.131993","volume":"591","author":"Y Zhang","year":"2025","unstructured":"Zhang, Y., Ju, X., Yan, C., Bo, J., Li, X., Zhang, J.: Polarized hyperspectral image fusion method for targets in sea clutter background. Opt. Commun. 591, 131993\u2013131993 (2025)","journal-title":"Opt. Commun."},{"key":"5083_CR10","doi-asserted-by":"crossref","unstructured":"Lin, A.Y., Turnbull, C.D., Pandit, J.J.: The \u201ctask\u201d of breathing: Anesthetic relevance of background two-pore domain potassium channels as therapeutic targets for respiratory control. Anesthesia and analgesia (2025)","DOI":"10.1213\/ANE.0000000000007365"},{"issue":"2","key":"5083_CR11","doi-asserted-by":"publisher","first-page":"269","DOI":"10.1016\/j.dcan.2023.11.001","volume":"11","author":"J Zhu","year":"2025","unstructured":"Zhu, J., Qin, C., Choi, D.: Yolo-sdluwd: Yolov7-based small target detection network for infrared images in complex backgrounds. Digital Communications and Networks 11(2), 269\u2013279 (2025)","journal-title":"Digital Communications and Networks"},{"issue":"18","key":"5083_CR12","doi-asserted-by":"publisher","first-page":"3492","DOI":"10.3390\/rs16183492","volume":"16","author":"H Chan","year":"2024","unstructured":"Chan, H., Qiu, X., Gao, X., Lu, D.: A complex background sar ship target detection method based on fusion tensor and cross-domain adversarial learning. Remote Sensing 16(18), 3492\u20133492 (2024)","journal-title":"Remote Sensing"},{"key":"5083_CR13","doi-asserted-by":"crossref","unstructured":"Chen, X., Yan, B., Zhu, J., Wang, D., Yang, X., Lu, H.: Transformer tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8126\u20138135 (2021)","DOI":"10.1109\/CVPR46437.2021.00803"},{"key":"5083_CR14","doi-asserted-by":"crossref","unstructured":"Cai, Y., Liu, J., Tang, J., Wu, G.: Robust object modeling for visual tracking. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9589\u20139600 (2023)","DOI":"10.1109\/ICCV51070.2023.00879"},{"key":"5083_CR15","doi-asserted-by":"crossref","unstructured":"Song, Z., Yu, J., Chen, Y.-P.P., Yang, W.: Transformer tracking with cyclic shifting window attention. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8791\u20138800 (2022)","DOI":"10.1109\/CVPR52688.2022.00859"},{"key":"5083_CR16","doi-asserted-by":"crossref","unstructured":"Gao, S., Zhou, C., Ma, C., Wang, X., Yuan, J.: Aiatrack: Attention in attention for transformer visual tracking. In: European Conference on Computer Vision, pp. 146\u2013164 . Springer (2022)","DOI":"10.1007\/978-3-031-20047-2_9"},{"key":"5083_CR17","first-page":"16743","volume":"35","author":"L Lin","year":"2022","unstructured":"Lin, L., Fan, H., Zhang, Z., Xu, Y., Ling, H.: Swintrack: a simple and strong baseline for transformer tracking. Adv. Neural. Inf. Process. Syst. 35, 16743\u201316754 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"5083_CR18","doi-asserted-by":"crossref","unstructured":"Yan, B., Peng, H., Fu, J., Wang, D., Lu, H.: Learning spatio-temporal transformer for visual tracking. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10448\u201310457 (2021)","DOI":"10.1109\/ICCV48922.2021.01028"},{"key":"5083_CR19","doi-asserted-by":"crossref","unstructured":"Wang, J., Lai, C., Wang, Y., Zhang, W.: Emat: efficient feature fusion network for visual tracking via optimized multi-head attention. Neural Networks: The Official Journal of the International Neural Network Society, 172 (2024)","DOI":"10.1016\/j.neunet.2024.106110"},{"issue":"000","key":"5083_CR20","first-page":"13","volume":"26","author":"J Wang","year":"2024","unstructured":"Wang, J., Yin, P., Wang, Y., Yang, W.: Cmat: integrating convolution mixer and self-attention for visual tracking. Multimedia, IEEE Trans. on (T-MM) 26(000), 13 (2024)","journal-title":"Multimedia, IEEE Trans. on (T-MM)"},{"key":"5083_CR21","doi-asserted-by":"crossref","unstructured":"Bhat, G., Danelljan, M., Gool, L.V., Timofte, R.: Learning discriminative model prediction for tracking. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6182\u20136191 (2019)","DOI":"10.1109\/ICCV.2019.00628"},{"key":"5083_CR22","doi-asserted-by":"crossref","unstructured":"Danelljan, M., Gool, L.V., Timofte, R.: Probabilistic regression for visual tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7183\u20137192 (2020)","DOI":"10.1109\/CVPR42600.2020.00721"},{"key":"5083_CR23","doi-asserted-by":"crossref","unstructured":"Fu, Z., Liu, Q., Fu, Z., Wang, Y.: Stmtrack: Template-free visual tracking with space-time memory networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13774\u201313783 (2021)","DOI":"10.1109\/CVPR46437.2021.01356"},{"key":"5083_CR24","doi-asserted-by":"crossref","unstructured":"Chen, B., Li, P., Bai, L., Qiao, L., Shen, Q., Li, B., Gan, W., Wu, W., Ouyang, W.: Backbone is all your need: A simplified architecture for visual object tracking. In: European Conference on Computer Vision, pp. 375\u2013392. Springer (2022)","DOI":"10.1007\/978-3-031-20047-2_22"},{"key":"5083_CR25","doi-asserted-by":"crossref","unstructured":"Cui, Y., Jiang, C., Wang, L., Wu, G.: Mixformer: End-to-end tracking with iterative mixed attention. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13608\u201313618 (2022)","DOI":"10.1109\/CVPR52688.2022.01324"},{"key":"5083_CR26","unstructured":"Zhou, X., Guo, P., Hong, L., Li, J., Zhang, W., Ge, W., Zhang, W.: Reading relevant feature from global representation memory for visual object tracking. Advances in Neural Information Processing Systems 36 (2024)"},{"key":"5083_CR27","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv:2010.11929 (2020)"},{"key":"5083_CR28","doi-asserted-by":"crossref","unstructured":"Ye, B., Chang, H., Ma, B., Shan, S., Chen, X.: Joint feature learning and relation modeling for tracking: A one-stream framework. In: European Conference on Computer Vision, pp. 341\u2013357. Springer (2022)","DOI":"10.1007\/978-3-031-20047-2_20"},{"key":"5083_CR29","doi-asserted-by":"crossref","unstructured":"Wu, Q., Yang, T., Liu, Z., Wu, B., Shan, Y., Chan, A.B.: Dropmae: Masked autoencoders with spatial-attention dropout for tracking tasks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14561\u201314571 (2023)","DOI":"10.1109\/CVPR52729.2023.01399"},{"issue":"5","key":"5083_CR30","doi-asserted-by":"publisher","first-page":"1562","DOI":"10.1109\/TPAMI.2019.2957464","volume":"43","author":"L Huang","year":"2019","unstructured":"Huang, L., Zhao, X., Huang, K.: Got-10k: a large high-diversity benchmark for generic object tracking in the wild. IEEE Trans. Pattern Anal. Mach. Intell. 43(5), 1562\u20131577 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"5083_CR31","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., Zitnick, C.L.: Microsoft coco: Common objects in context. In: Computer Vision\u2013ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, pp. 740\u2013755. Springer (2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"5083_CR32","doi-asserted-by":"crossref","unstructured":"Fan, H., Bai, H., Lin, L., Yang, F., Chu, P., Deng, G., Yu, S., Harshit, Huang, M., Liu, J., Lasot: A high-quality large-scale single object tracking benchmark, et al.: Lasot: a high-quality large-scale single object tracking benchmark. Int. J. Comput. Vis. 129, 439\u2013461 (2021)","DOI":"10.1007\/s11263-020-01387-y"},{"key":"5083_CR33","doi-asserted-by":"crossref","unstructured":"Muller, M., Bibi, A., Giancola, S., Alsubaihi, S., Ghanem, B.: Trackingnet: A large-scale dataset and benchmark for object tracking in the wild. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 300\u2013317 (2018)","DOI":"10.1007\/978-3-030-01246-5_19"},{"key":"5083_CR34","doi-asserted-by":"crossref","unstructured":"Fan, H., Lin, L., Yang, F., Chu, P., Deng, G., Yu, S., Bai, H., Xu, Y., Liao, C., Ling, H.: Lasot: A high-quality benchmark for large-scale single object tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5374\u20135383 (2019)","DOI":"10.1109\/CVPR.2019.00552"},{"key":"5083_CR35","doi-asserted-by":"crossref","unstructured":"Xie, J., Zhong, B., Mo, Z., Zhang, S., Shi, L., Song, S., Ji, R.: Autoregressive queries for adaptive tracking with spatio-temporal transformers. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 19300\u201319309 (2024)","DOI":"10.1109\/CVPR52733.2024.01826"},{"key":"5083_CR36","doi-asserted-by":"crossref","unstructured":"Gao, S., Zhou, C., Zhang, J.: Generalized relation modeling for transformer tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 18686\u201318695 (2023)","DOI":"10.1109\/CVPR52729.2023.01792"},{"key":"5083_CR37","doi-asserted-by":"crossref","unstructured":"Bhat, G., Danelljan, M., Van\u00a0Gool, L., Timofte, R.: Know your surroundings: Exploiting scene information for object tracking. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XXIII 16, pp. 205\u2013221. Springer (2020)","DOI":"10.1007\/978-3-030-58592-1_13"},{"key":"5083_CR38","doi-asserted-by":"crossref","unstructured":"Mayer, C., Danelljan, M., Paudel, D.P., Van\u00a0Gool, L.: Learning target candidate association to keep track of what not to track. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 13444\u201313454 (2021)","DOI":"10.1109\/ICCV48922.2021.01319"},{"key":"5083_CR39","doi-asserted-by":"crossref","unstructured":"Kiani\u00a0Galoogahi, H., Fagg, A., Huang, C., Ramanan, D., Lucey, S.: Need for speed: A benchmark for higher frame rate object tracking. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1125\u20131134 (2017)","DOI":"10.1109\/ICCV.2017.128"},{"issue":"9","key":"5083_CR40","doi-asserted-by":"publisher","first-page":"1834","DOI":"10.1109\/TPAMI.2014.2388226","volume":"37","author":"Y Wu","year":"2015","unstructured":"Wu, Y., Lim, J., Yang, M.-H.: Object tracking benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 37(9), 1834\u20131848 (2015). https:\/\/doi.org\/10.1109\/TPAMI.2014.2388226","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"5083_CR41","unstructured":"Benchmark, U.: A benchmark and simulator for uav tracking. In: European Conference on Computer Vision, vol. 7 (2016)"},{"key":"5083_CR42","doi-asserted-by":"crossref","unstructured":"Li, X., Ma, C., Wu, B., He, Z., Yang, M.-H.: Target-aware deep tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1369\u20131378 (2019)","DOI":"10.1109\/CVPR.2019.00146"},{"key":"5083_CR43","doi-asserted-by":"crossref","unstructured":"Danelljan, M., Bhat, G., Khan, F.S., Felsberg, M.: Atom: Accurate tracking by overlap maximization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4660\u20134669 (2019)","DOI":"10.1109\/CVPR.2019.00479"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-05083-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-025-05083-7","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-05083-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,26]],"date-time":"2026-01-26T15:44:01Z","timestamp":1769442241000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-025-05083-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1]]},"references-count":43,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,1]]}},"alternative-id":["5083"],"URL":"https:\/\/doi.org\/10.1007\/s11760-025-05083-7","relation":{"has-preprint":[{"id-type":"doi","id":"10.21203\/rs.3.rs-7288675\/v1","asserted-by":"object"}]},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"value":"1863-1703","type":"print"},{"value":"1863-1711","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1]]},"assertion":[{"value":"4 August 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 November 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 December 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 January 2026","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"We declare that we have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest"}},{"value":"The authors declare no competing interests.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}},{"value":"This declaration is not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}],"article-number":"35"}}