{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,9]],"date-time":"2026-01-09T23:54:43Z","timestamp":1768002883804,"version":"3.49.0"},"reference-count":41,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2024,12,19]],"date-time":"2024-12-19T00:00:00Z","timestamp":1734566400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2024,12,19]],"date-time":"2024-12-19T00:00:00Z","timestamp":1734566400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"name":"Key Research and Development Program of Linyi City","award":["2022028"],"award-info":[{"award-number":["2022028"]}]},{"DOI":"10.13039\/501100007129","name":"Natural Science Foundation of Shandong Province","doi-asserted-by":"publisher","award":["ZR2019MA030"],"award-info":[{"award-number":["ZR2019MA030"]}],"id":[{"id":"10.13039\/501100007129","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2025,1]]},"DOI":"10.1007\/s40747-024-01672-0","type":"journal-article","created":{"date-parts":[[2024,12,19]],"date-time":"2024-12-19T09:48:52Z","timestamp":1734601732000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["ADSTrack: adaptive dynamic sampling for visual tracking"],"prefix":"10.1007","volume":"11","author":[{"given":"Zhenhai","family":"Wang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0097-3318","authenticated-orcid":false,"given":"Lutao","family":"Yuan","sequence":"additional","affiliation":[]},{"given":"Ying","family":"Ren","sequence":"additional","affiliation":[]},{"given":"Sen","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Hongyu","family":"Tian","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,19]]},"reference":[{"key":"1672_CR1","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez A. N, Kaiser \u0141, Polosukhin I (2017), Attention is all you need, Advances in neural information processing systems, 30"},{"key":"1672_CR2","doi-asserted-by":"crossref","unstructured":"Chen X, Yan B, Zhu J, Wang D, Yang X, Lu H (2021), Transformer tracking, in Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 8126\u20138135","DOI":"10.1109\/CVPR46437.2021.00803"},{"key":"1672_CR3","doi-asserted-by":"crossref","unstructured":"Yan B, Peng H, Fu J, Wang D, Lu H (2021), Learning spatio-temporal transformer for visual tracking, in Proceedings of the IEEE\/CVF international conference on computer vision, 10448\u201310457","DOI":"10.1109\/ICCV48922.2021.01028"},{"key":"1672_CR4","doi-asserted-by":"crossref","unstructured":"Wang N, Zhou W, Wang J, Li H (2021), Transformer meets tracker: Exploiting temporal context for robust visual tracking, in Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 1571\u20131580","DOI":"10.1109\/CVPR46437.2021.00162"},{"key":"1672_CR5","doi-asserted-by":"crossref","unstructured":"Ye B, Chang H, Ma B, Shan S, Chen X (2022), Joint feature learning and relation modeling for tracking: A one-stream framework, in European Conference on Computer Vision, 341\u2013357, Springer","DOI":"10.1007\/978-3-031-20047-2_20"},{"key":"1672_CR6","doi-asserted-by":"crossref","unstructured":"Gao S, Zhou C, Ma C, Wang X, Yuan J (2022), Aiatrack: Attention in attention for transformer visual tracking, in European Conference on Computer Vision,146\u2013164, Springer","DOI":"10.1007\/978-3-031-20047-2_9"},{"key":"1672_CR7","doi-asserted-by":"crossref","unstructured":"Xie F, Wang C, Wang G, Cao Y, Yang W, Zeng W (2022), Correlation-aware deep tracking, in Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 8751\u20138760","DOI":"10.1109\/CVPR52688.2022.00855"},{"key":"1672_CR8","doi-asserted-by":"crossref","unstructured":"Lan J.-P, Cheng Z.-Q, He J.-Y, Li C, Luo B, Bao X, Xiang W, Geng Y, Xie X (2023), Procontext: Exploring progressive context transformer for tracking, in ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 1\u20135, IEEE","DOI":"10.1109\/ICASSP49357.2023.10094971"},{"key":"1672_CR9","doi-asserted-by":"crossref","unstructured":"Cao Z, Fu C, Ye J, Li B, Li Y (2021), Hift: Hierarchical feature transformer for aerial tracking, in Proceedings of the IEEE\/CVF International Conference on Computer Vision, 15457\u201315466","DOI":"10.1109\/ICCV48922.2021.01517"},{"key":"1672_CR10","unstructured":"Krizhevsky A, Sutskever I, Hinton G. E (2012), Imagenet classification with deep convolutional neural networks, Advances in neural information processing systems, 25"},{"key":"1672_CR11","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016), Deep residual learning for image recognition, in Proceedings of the IEEE conference on computer vision and pattern recognition, 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"1672_CR12","doi-asserted-by":"crossref","unstructured":"Cui Y, Jiang C, Wang L, Wu G (2022), Mixformer: End-to-end tracking with iterative mixed attention, in Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 13608\u201313618","DOI":"10.1109\/CVPR52688.2022.01324"},{"key":"1672_CR13","first-page":"16743","volume":"35","author":"L Lin","year":"2022","unstructured":"Lin L, Fan H, Zhang Z, Xu Y, Ling H (2022) Swintrack: A simple and strong baseline for transformer tracking. Adv Neural Inf Process Syst 35:16743\u201316754","journal-title":"Adv Neural Inf Process Syst"},{"key":"1672_CR14","doi-asserted-by":"crossref","unstructured":"Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, Lin S, Guo B (2021), Swin transformer: Hierarchical vision transformer using shifted windows, in Proceedings of the IEEE\/CVF international conference on computer vision, 10012\u201310022","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"1672_CR15","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S (2020), et al., An image is worth 16x16 words: Transformers for image recognition at scale, arXiv preprint arXiv:2010.11929"},{"key":"1672_CR16","first-page":"13937","volume":"34","author":"Y Rao","year":"2021","unstructured":"Rao Y, Zhao W, Liu B, Lu J, Zhou J, Hsieh C-J (2021) Dynamicvit: Efficient vision transformers with dynamic token sparsification. Adv Neural Inf Process Syst 34:13937\u201313949","journal-title":"Adv Neural Inf Process Syst"},{"key":"1672_CR17","doi-asserted-by":"crossref","unstructured":"Fayyaz M, Koohpayegani S. A, Jafari F. R, Sengupta S, Joze H. R. V, Sommerlade E, Pirsiavash H, Gall J (2022), Adaptive token sampling for efficient vision transformers, in European Conference on Computer Vision, 396\u2013414, Springer","DOI":"10.1007\/978-3-031-20083-0_24"},{"key":"1672_CR18","doi-asserted-by":"crossref","unstructured":"Yin H, Vahdat A, Alvarez J. M, Mallya A, Kautz J, Molchanov P (2022), A-vit: Adaptive tokens for efficient vision transformer, in Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 10809\u201310818","DOI":"10.1109\/CVPR52688.2022.01054"},{"key":"1672_CR19","unstructured":"Liang Y, Ge C, Tong Z, Song Y, Wang J, Xie P (2022), Not all patches are what you need: Expediting vision transformers via token reorganizations, arXiv preprint arXiv:2202.07800"},{"key":"1672_CR20","doi-asserted-by":"crossref","unstructured":"Zhou Q, Zhu Y (2023), Make a long image short: Adaptive token length for vision transformers, in Joint European Conference on Machine Learning and Knowledge Discovery in Databases, 69\u201385, Springer","DOI":"10.1007\/978-3-031-43415-0_5"},{"key":"1672_CR21","unstructured":"Darcet T, Oquab M, Mairal J, Bojanowski P (2023), Vision transformers need registers, arXiv preprint arXiv:2309.16588"},{"key":"1672_CR22","doi-asserted-by":"crossref","unstructured":"He K, Chen X, Xie S, Li Y, Doll\u00e1r P, Girshick R (2022), Masked autoencoders are scalable vision learners, in Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, 16000\u201316009","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"1672_CR23","unstructured":"Zhou X, Wang D, Kr\u00e4henb\u00fchl P (2019), Objects as points, arXiv preprint arXiv:1904.07850"},{"key":"1672_CR24","doi-asserted-by":"crossref","unstructured":"Lin T.-Y, Maire M, Belongie S, Hays J, Perona P, Ramanan D, Doll\u00e1r P, Zitnick C. L (2014), Microsoft coco: Common objects in context, in European conference on computer vision, 740\u2013755, Springer","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"1672_CR25","doi-asserted-by":"crossref","unstructured":"Fan H, Lin L, Yang F, Chu P, Deng G, Yu S, Bai H, Xu Y, Liao C, Ling H (2019), Lasot: A high-quality benchmark for large-scale single object tracking, in Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, 5374\u20135383","DOI":"10.1109\/CVPR.2019.00552"},{"issue":"5","key":"1672_CR26","doi-asserted-by":"publisher","first-page":"1562","DOI":"10.1109\/TPAMI.2019.2957464","volume":"43","author":"L Huang","year":"2019","unstructured":"Huang L, Zhao X, Huang K (2019) Got-10k: A large high-diversity benchmark for generic object tracking in the wild. IEEE Trans Pattern Anal Mach Intell 43(5):1562\u20131577","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1672_CR27","doi-asserted-by":"crossref","unstructured":"Muller M, Bibi A, Giancola S, Alsubaihi S, Ghanem B (2018), Trackingnet: A large-scale dataset and benchmark for object tracking in the wild, in Proceedings of the European conference on computer vision (ECCV), 300\u2013317","DOI":"10.1007\/978-3-030-01246-5_19"},{"key":"1672_CR28","unstructured":"Loshchilov I, Hutter F (2017), Decoupled weight decay regularization, arXiv preprint arXiv:1711.05101"},{"issue":"37","key":"1672_CR29","doi-asserted-by":"publisher","first-page":"1834","DOI":"10.1109\/TPAMI.2014.2388226","volume":"9","author":"Y Wu","year":"2015","unstructured":"Wu Y, Lim J, Yang M-H (2015) Object tracking benchmark. IEEE Trans Pattern Anal Mach Intell 9(37):1834\u20131848","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1672_CR30","doi-asserted-by":"crossref","unstructured":"Kiani Galoogahi H, Fagg A, Huang C, Ramanan D, Lucey S (2017), Need for speed: A benchmark for higher frame rate object tracking, in Proceedings of the IEEE International Conference on Computer Vision, 1125\u20131134","DOI":"10.1109\/ICCV.2017.128"},{"key":"1672_CR31","doi-asserted-by":"crossref","unstructured":"Mueller M, Smith N, Ghanem B (2016), A benchmark and simulator for uav tracking, in European conference on computer vision, 445\u2013461, Springer","DOI":"10.1007\/978-3-319-46448-0_27"},{"key":"1672_CR32","doi-asserted-by":"crossref","unstructured":"Bertinetto L, Valmadre J, Henriques JF, Vedaldi A, Torr PH (2016) Fully-convolutional siamese networks for object tracking, in Computer Vision-ECCV 2016 Workshops: Amsterdam, The Netherlands, October 8\u201310 and 15\u201316, 2016, Proceedings, Part II 14, 850\u2013865. Springer","DOI":"10.1007\/978-3-319-48881-3_56"},{"key":"1672_CR33","doi-asserted-by":"crossref","unstructured":"Danelljan M, Bhat G, Shahbaz Khan F, Felsberg M (2017), Eco: Efficient convolution operators for tracking, in Proceedings of the IEEE conference on computer vision and pattern recognition, 6638\u20136646","DOI":"10.1109\/CVPR.2017.733"},{"key":"1672_CR34","doi-asserted-by":"crossref","unstructured":"Zhang Z, Peng H, Fu J, Li B, Hu W (2020), Ocean: Object-aware anchor-free tracking, in European Conference on Computer Vision, 771\u2013787, Springer","DOI":"10.1007\/978-3-030-58589-1_46"},{"key":"1672_CR35","doi-asserted-by":"crossref","unstructured":"Li B, Wu W, Wang Q, Zhang F, Xing J, Yan J (2019), Siamrpn++: Evolution of siamese visual tracking with very deep networks, in Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 4282\u20134291","DOI":"10.1109\/CVPR.2019.00441"},{"key":"1672_CR36","doi-asserted-by":"crossref","unstructured":"Chen Z, Zhong B, Li G, Zhang S, Ji R (2020), Siamese box adaptive network for visual tracking, in Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, 6668\u20136677","DOI":"10.1109\/CVPR42600.2020.00670"},{"key":"1672_CR37","doi-asserted-by":"crossref","unstructured":"Danelljan M, Bhat G, Khan F. S, Felsberg M (2019), Atom: Accurate tracking by overlap maximization, in Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 4660\u20134669,","DOI":"10.1109\/CVPR.2019.00479"},{"key":"1672_CR38","doi-asserted-by":"crossref","unstructured":"Bhat G, Danelljan M, Gool L. V, Timofte R (2019), Learning discriminative model prediction for tracking, in Proceedings of the IEEE\/CVF international conference on computer vision, 6182\u20136191","DOI":"10.1109\/ICCV.2019.00628"},{"key":"1672_CR39","doi-asserted-by":"crossref","unstructured":"Fu Z,Liu Q, Fu Z, Wang Y (2021), Stmtrack: Template-free visual tracking with space-time memory networks, in Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, 13774\u201313783","DOI":"10.1109\/CVPR46437.2021.01356"},{"key":"1672_CR40","doi-asserted-by":"crossref","unstructured":"Pi Z,Wan W, Sun C, Gao C, Sang N, Li C (2022), Hierarchical feature embedding for visual tracking, in European Conference on Computer Vision, 428\u2013445, Springer","DOI":"10.1007\/978-3-031-20047-2_25"},{"key":"1672_CR41","doi-asserted-by":"publisher","first-page":"1549","DOI":"10.1609\/aaai.v37i2.25241","volume":"37","author":"Y Liang","year":"2023","unstructured":"Liang Y, Li Q, Long F (2023) Global dilated attention and target focusing network for robust tracking. Proceedings of the AAAI Conference on Artificial Intelligence 37:1549\u20131557","journal-title":"Proceedings of the AAAI Conference on Artificial Intelligence"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01672-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-024-01672-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01672-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,30]],"date-time":"2025-01-30T20:17:29Z","timestamp":1738268249000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-024-01672-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,19]]},"references-count":41,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,1]]}},"alternative-id":["1672"],"URL":"https:\/\/doi.org\/10.1007\/s40747-024-01672-0","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,19]]},"assertion":[{"value":"3 March 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 November 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 December 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"We confirm that the manuscript has been read and approved by all named authors. We understand that the corresponding author is the sole contact for the editorial process. He is responsible for communicating with the other authors about progress, submissions of revisions and final approval of proofs.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Author agreement"}}],"article-number":"79"}}