{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,5]],"date-time":"2025-11-05T21:12:58Z","timestamp":1762377178032,"version":"3.37.3"},"reference-count":51,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2022,5,31]],"date-time":"2022-05-31T00:00:00Z","timestamp":1653955200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2022,5,31]],"date-time":"2022-05-31T00:00:00Z","timestamp":1653955200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"name":"the National Science Fund for Distinguished Young Scholars of China","award":["51625501"],"award-info":[{"award-number":["51625501"]}]},{"DOI":"10.13039\/501100004750","name":"Aeronautical Science Foundation of China","doi-asserted-by":"publisher","award":["201946051002"],"award-info":[{"award-number":["201946051002"]}],"id":[{"id":"10.13039\/501100004750","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52127809"],"award-info":[{"award-number":["52127809"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2023,2]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Visual tracking is a crucial task in computer vision that concentrates on assessing the state of the target in each frame. Siamese based trackers possess an excellent tracking mechanism and balance the accuracy and efficiency well through continuous optimization of network structure. However, the rapid appearance variation and occlusion are still huge challenges to the accuracy and success rate of tracking task. Most Siamese based trackers rely on a fixed object template to match the target in search area and also neglect the importance of feature representation to tracking tasks. Based on the core idea of how to promote the recognition ability of trackers for the dynamic object and keep stable tracking during occlusion process, we present Siamese block attention network for online update object tracking referred to as SBAN. The proposed Siamese block attention module adopts a Siamese network structure to integrate two complementary global descriptors and establish the interdependence among channels to generate channel weights which can enhance the crucial features and restrain inessential ones. We also design an online update target module that can effectively utilize the history tracking information. The final updated target module is the integration of the given template, the process template, and the last tracking results. Experiments on four benchmarks, OTB2015, VOT2018, UAV123, LaSOT, illustrate that our tracker obtains outstanding tracking performance in accuracy and robustness while running at over 53 FPS.<\/jats:p>","DOI":"10.1007\/s10489-022-03619-9","type":"journal-article","created":{"date-parts":[[2022,5,31]],"date-time":"2022-05-31T20:02:50Z","timestamp":1654027370000},"page":"3459-3471","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["Siamese block attention network for online update object tracking"],"prefix":"10.1007","volume":"53","author":[{"given":"Dingkun","family":"Xiao","sequence":"first","affiliation":[]},{"given":"Ke","family":"Tan","sequence":"additional","affiliation":[]},{"given":"Zhenzhong","family":"Wei","sequence":"additional","affiliation":[]},{"given":"Guangjun","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,5,31]]},"reference":[{"issue":"1","key":"3619_CR1","first-page":"61","volume":"31","author":"HC Lu","year":"2018","unstructured":"Lu HC, Li PX, Wang D (2018) Visual object tracking: a survey. Pattern Recognit Artif Intell 31(1):61\u201376","journal-title":"Pattern Recognit Artif Intell"},{"key":"3619_CR2","doi-asserted-by":"crossref","unstructured":"Tang S, Andriluka M, Andres B et al (2017) Multiple people tracking by lifted multicut and person re-identification. Proc IEEE Conf Comput Vis Pattern Recognit:3539\u20133548","DOI":"10.1109\/CVPR.2017.394"},{"key":"3619_CR3","doi-asserted-by":"publisher","first-page":"103448","DOI":"10.1016\/j.artint.2020.103448","volume":"293","author":"W Luo","year":"2021","unstructured":"Luo W, Xing J, Milan A, Zhang X, Liu W, Kim TK (2021) Multiple object tracking: a literature review. Artif Intell 293:103448","journal-title":"Artif Intell"},{"key":"3619_CR4","doi-asserted-by":"crossref","unstructured":"Marvasti-Zadeh SM, Cheng L, Ghanei-Yakhdan H, Kasaei S (2021) Deep learning for visual tracking: A comprehensive survey. IEEE Transactions on Intelligent Transportation Systems","DOI":"10.1109\/TITS.2020.3046478"},{"key":"3619_CR5","doi-asserted-by":"publisher","first-page":"61","DOI":"10.1016\/j.neucom.2019.11.023","volume":"381","author":"G Ciaparrone","year":"2020","unstructured":"Ciaparrone G, S\u00e1nchez FL, Tabik S et al (2020) Deep learning in video multi-object tracking: a survey. Neurocomputing 381:61\u201388","journal-title":"Neurocomputing"},{"key":"3619_CR6","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1016\/j.patcog.2017.11.007","volume":"76","author":"P Li","year":"2018","unstructured":"Li P, Wang D, Wang L, Lu H (2018) Deep visual tracking: review and experimental comparison. Pattern Recogn 76:323\u2013338","journal-title":"Pattern Recogn"},{"key":"3619_CR7","doi-asserted-by":"crossref","unstructured":"Bertinetto L, Valmadre J, Henriques JF, Vedaldi A, Torr PH (2016). Fully-convolutional siamese networks for object tracking. In Eur Conf Comput Vis.\u00a0Springer, Cham pp 850\u2013865","DOI":"10.1007\/978-3-319-48881-3_56"},{"key":"3619_CR8","doi-asserted-by":"crossref","unstructured":"Li B, Yan J, Wu W et al (2018) High performance visual tracking with siamese region proposal network. Proc IEEE Conf Comput Vis Pattern Recognit:8971\u20138980","DOI":"10.1109\/CVPR.2018.00935"},{"key":"3619_CR9","doi-asserted-by":"crossref","unstructured":"Li B, Wu W, Wang Q et al (2019) Siamrpn++: evolution of siamese visual tracking with very deep networks. Proceedings of the IEEE\/CVF Conf Comput Vis Pattern Recognit:4282\u20134291","DOI":"10.1109\/CVPR.2019.00441"},{"issue":"9","key":"3619_CR10","doi-asserted-by":"publisher","first-page":"1834","DOI":"10.1109\/TPAMI.2014.2388226","volume":"37","author":"Y Wu","year":"2015","unstructured":"Wu Y, Lim J, Yang MH (2015) Object tracking benchmark. IEEE Trans Pattern Anal Machine Intell 37(9):1834\u20131848","journal-title":"IEEE Trans Pattern Anal Machine Intell"},{"key":"3619_CR11","unstructured":"Kristan M, Leonardis A, Matas J, Felsberg M, Pflugfelder R, \u02c7Cehovin Zajc L, ... Sun Y (2018). The sixth visual object tracking vot2018 challenge results. In\u00a0Proc Eur Conf Comput Vis\u00a0(ECCV) Workshops pp 0\u20130"},{"key":"3619_CR12","doi-asserted-by":"crossref","unstructured":"Mueller M, Smith N, Ghanem B (2016). A benchmark and simulator for uav tracking. In Eur Conf Comput Vis.\u00a0Springer, Cham pp 445\u2013461\u00a0","DOI":"10.1007\/978-3-319-46448-0_27"},{"key":"3619_CR13","doi-asserted-by":"crossref","unstructured":"Fan H, Lin L, Yang F et al (2019) Lasot: a high-quality benchmark for large-scale single object tracking. Proceedings of the IEEE\/CVF Conference Comput Vis Pattern Recognit 5374\u20135383","DOI":"10.1109\/CVPR.2019.00552"},{"key":"3619_CR14","doi-asserted-by":"publisher","first-page":"1895","DOI":"10.1007\/s40747-020-00161-4","volume":"7","author":"S Liu","year":"2021","unstructured":"Liu S, Liu D, Srivastava G, Po\u0142ap D, Wo\u017aniak M (2021) Overview and methods of correlation filter algorithms in object tracking. Complex Intell Syst 7:1895\u20131917","journal-title":"Complex Intell Syst"},{"key":"3619_CR15","doi-asserted-by":"crossref","unstructured":"Bolme DS, Beveridge JR, Draper BA, Lui YM (2010). Visual object tracking using adaptive correlation filters. In 2010 IEEE Comput Soc Conf Comput Vis Pattern Recognit.\u00a0IEEE pp 2544-2550","DOI":"10.1109\/CVPR.2010.5539960"},{"issue":"3","key":"3619_CR16","doi-asserted-by":"publisher","first-page":"583","DOI":"10.1109\/TPAMI.2014.2345390","volume":"37","author":"JF Henriques","year":"2015","unstructured":"Henriques JF, Caseiro R, Martins P, Batista J (2015) High-speed tracking with kernelized correlation filters. IEEE Trans Pattern Anal Machine Intell 37(3):583\u2013596","journal-title":"IEEE Trans Pattern Anal Machine Intell"},{"key":"3619_CR17","doi-asserted-by":"crossref","unstructured":"Danelljan M, Bhat G, Shahbaz Khan F et al (2017) Eco: efficient convolution operators for tracking. Proc IEEE Conf Comput Vis Pattern Recognit:6638\u20136646","DOI":"10.1109\/CVPR.2017.733"},{"issue":"5","key":"3619_CR18","doi-asserted-by":"publisher","first-page":"2526","DOI":"10.1109\/TIP.2018.2806280","volume":"27","author":"E Gundogdu","year":"2018","unstructured":"Gundogdu E, Alatan AA (2018) Good features to correlate for visual tracking. IEEE Trans Image Process 27(5):2526\u20132540","journal-title":"IEEE Trans Image Process"},{"issue":"4","key":"3619_CR19","doi-asserted-by":"publisher","first-page":"1268","DOI":"10.1109\/TCSVT.2019.2944654","volume":"31","author":"X Lu","year":"2019","unstructured":"Lu X, Ma C, Ni B et al (2019) Adaptive region proposal with channel regularization for robust object tracking. IEEE Trans Circuits Systems vid Technol 31(4):1268\u20131282","journal-title":"IEEE Trans Circuits Systems vid Technol"},{"issue":"12","key":"3619_CR20","doi-asserted-by":"publisher","first-page":"4855","DOI":"10.1007\/s12652-018-01171-4","volume":"10","author":"Y Chen","year":"2019","unstructured":"Chen Y, Wang J, Xia R, Zhang Q, Cao Z, Yang K (2019) RETRACTED ARTICLE: the visual object tracking algorithm research based on adaptive combination kernel. J Ambient Intell Humaniz Comput 10(12):4855\u20134867","journal-title":"J Ambient Intell Humaniz Comput"},{"key":"3619_CR21","doi-asserted-by":"publisher","first-page":"9614","DOI":"10.1109\/TIP.2020.3029897","volume":"29","author":"Y Zheng","year":"2020","unstructured":"Zheng Y, Liu X, Cheng X, Zhang K, Wu Y, Chen S (2020) Multi-task deep dual correlation filters for visual tracking. IEEE Trans Image Process 29:9614\u20139626","journal-title":"IEEE Trans Image Process"},{"key":"3619_CR22","doi-asserted-by":"crossref","unstructured":"Guo Q, Feng W, Zhou C, Huang R, Wan L, Wang S (2017). Learning dynamic siamese network for visual object tracking. In Proc IEEE Int Conf Comput Vis pp 1763\u20131771","DOI":"10.1109\/ICCV.2017.196"},{"key":"3619_CR23","doi-asserted-by":"crossref","unstructured":"Zhu Z, Wang Q, Li B, Wu W, Yan J, Hu W (2018). Distractor-aware siamese networks for visual object tracking. In Proc Eur Conf Comput Vis (ECCV) pp 101\u2013117","DOI":"10.1007\/978-3-030-01240-3_7"},{"key":"3619_CR24","doi-asserted-by":"crossref","unstructured":"Voigtlaender P, Luiten J, Torr PH, Leibe B (2020). Siam r-cnn: Visual tracking by re-detection. In Proc IEEE\/CVF Conf Comput Vis Pattern Recognit pp 6578\u20136588","DOI":"10.1109\/CVPR42600.2020.00661"},{"key":"3619_CR25","doi-asserted-by":"publisher","first-page":"107170","DOI":"10.1016\/j.patcog.2019.107170","volume":"100","author":"Y Xiao","year":"2020","unstructured":"Xiao Y, Li J, Du B et al (2020) MeMu: metric correlation Siamese network and multi-class negative sampling for visual tracking. Pattern Recogn 100:107170","journal-title":"Pattern Recogn"},{"key":"3619_CR26","doi-asserted-by":"publisher","first-page":"4631","DOI":"10.1007\/s10489-020-01783-4","volume":"50","author":"T Li","year":"2020","unstructured":"Li T, Wu P, Ding F, Yang W (2020) Parallel dual networks for visual object tracking. Appl Intell 50:4631\u20134646","journal-title":"Appl Intell"},{"issue":"07","key":"3619_CR27","doi-asserted-by":"publisher","first-page":"12549","DOI":"10.1609\/aaai.v34i07.6944","volume":"34","author":"Y Xu","year":"2020","unstructured":"Xu Y, Wang Z, Li Z, Yuan Y, Yu G (2020) Siamfc++: towards robust and accurate visual tracking with target estimation guidelines. Proceedings of the AAAI Conference on Artificial Intelligence 34(07):12549\u201312556","journal-title":"Proceedings of the AAAI Conference on Artificial Intelligence"},{"key":"3619_CR28","doi-asserted-by":"crossref","unstructured":"Chen Z, Zhong B, Li G et al (2020) Siamese box adaptive network for visual tracking. Proceedings of the IEEE\/CVF Conf Comput Vis Pattern Recognit:6668\u20136677","DOI":"10.1109\/CVPR42600.2020.00670"},{"key":"3619_CR29","doi-asserted-by":"crossref","unstructured":"Guo D, Wang J, Cui Y et al (2020) SiamCAR: Siamese fully convolutional classification and regression for visual tracking. Proceedings of the IEEE\/CVF Conf Comput Vis Pattern Recognit:6269\u20136277","DOI":"10.1109\/CVPR42600.2020.00630"},{"key":"3619_CR30","doi-asserted-by":"crossref","unstructured":"Zhou Y, Zhang Y (2022) SiamET: a Siamese based visual tracking network with enhanced templates. Appl Intell","DOI":"10.1007\/s10489-021-03057-z"},{"key":"3619_CR31","doi-asserted-by":"crossref","unstructured":"Choi J, Kwon J, Lee KM (2019). Deep meta learning for real-time target-aware visual tracking. In Proc IEEE\/CVF Int Conf Comput Vis pp 911\u2013920","DOI":"10.1109\/ICCV.2019.00100"},{"key":"3619_CR32","doi-asserted-by":"crossref","unstructured":"Zhang L, Gonzalez-Garcia A, Weijer J et al (2019) Learning the model update for siamese trackers. Proceedings of the IEEE\/CVF Int Conf Comput Vis:4010\u20134019","DOI":"10.1109\/ICCV.2019.00411"},{"key":"3619_CR33","doi-asserted-by":"crossref","unstructured":"Li P, Chen B, Ouyang W et al (2019) Gradnet: gradient-guided network for visual object tracking. Proceedings of the IEEE\/CVF Int Conf Comput Vis:6162\u20136171","DOI":"10.1109\/ICCV.2019.00626"},{"key":"3619_CR34","doi-asserted-by":"publisher","first-page":"52","DOI":"10.1016\/j.ins.2019.12.084","volume":"517","author":"P Gao","year":"2020","unstructured":"Gao P, Zhang Q, Wang F, Xiao L, Fujita H, Zhang Y (2020) Learning reinforced attentional representation for end-to-end visual tracking. Inf Sci 517:52\u201367","journal-title":"Inf Sci"},{"key":"3619_CR35","doi-asserted-by":"crossref","unstructured":"Danelljan M, Bhat G, Khan FS, Felsberg M (2019). Atom: Accurate tracking by overlap maximization. In Proc IEEE\/CVF Conf Comput Vis Pattern Recognit pp 4660\u20134669","DOI":"10.1109\/CVPR.2019.00479"},{"key":"3619_CR36","doi-asserted-by":"crossref","unstructured":"Bhat G, Danelljan M, Gool LV, Timofte R (2019). Learning discriminative model prediction for tracking. In Proc IEEE\/CVF Int Conf Comput Vision pp 6182\u20136191","DOI":"10.1109\/ICCV.2019.00628"},{"key":"3619_CR37","doi-asserted-by":"crossref","unstructured":"Wang Q, Teng Z, Xing J et al (2018) Learning attentions: residual attentional siamese network for high performance online visual tracking. Proc IEEE Conf Comput Vis Pattern Recognit:4854\u20134863","DOI":"10.1109\/CVPR.2018.00510"},{"key":"3619_CR38","doi-asserted-by":"crossref","unstructured":"Yu Y, Xiong Y, Huang W et al (2020) Deformable siamese attention networks for visual object tracking. Proceedings of the IEEE\/CVF Conf Comput Vis Pattern Recognit:6728\u20136737","DOI":"10.1109\/CVPR42600.2020.00676"},{"key":"3619_CR39","doi-asserted-by":"publisher","first-page":"80","DOI":"10.1016\/j.patcog.2018.10.005","volume":"87","author":"B Chen","year":"2019","unstructured":"Chen B, Li P, Sun C, Wang D, Yang G, Lu H (2019) Multi attention module for visual tracking. Pattern Recogn 87:80\u201393","journal-title":"Pattern Recogn"},{"key":"3619_CR40","doi-asserted-by":"publisher","first-page":"8230","DOI":"10.1007\/s10489-021-02234-4","volume":"51","author":"Z Cui","year":"2021","unstructured":"Cui Z, Lu N (2021) Feature selection accelerated convolutional neural networks for visual tracking. Appl Intell 51:8230\u20138244","journal-title":"Appl Intell"},{"key":"3619_CR41","doi-asserted-by":"crossref","unstructured":"Chen X, Yan B, Zhu J et al (2021) Transformer tracking. Proceedings of the IEEE\/CVF Conf Comput Vis Pattern Recognit:8126\u20138135","DOI":"10.1109\/CVPR46437.2021.00803"},{"key":"3619_CR42","doi-asserted-by":"crossref","unstructured":"Wang N, Zhou W, Wang J et al (2021) Transformer meets tracker: exploiting temporal context for robust visual tracking. Proceedings of the IEEE\/CVF Conf Comput Vis Pattern Recognit:1571\u20131580","DOI":"10.1109\/CVPR46437.2021.00162"},{"key":"3619_CR43","unstructured":"Vaswani A, Shazeer N, Parmar N et al (2017) Attention is all you need. Adv Neural Inf Proces Syst 30"},{"issue":"11","key":"3619_CR44","doi-asserted-by":"publisher","first-page":"2709","DOI":"10.1109\/TPAMI.2018.2865311","volume":"41","author":"C Ma","year":"2019","unstructured":"Ma C, Huang J, Yang X, Yang M (2019) Robust Visual Tracking via Hierarchical Convolutional Features. IEEE Trans Pattern Anal Machine Intell 41(11):2709\u20132723","journal-title":"IEEE Trans Pattern Anal Machine Intell"},{"key":"3619_CR45","doi-asserted-by":"publisher","first-page":"235","DOI":"10.1016\/j.neucom.2016.12.002","volume":"237","author":"W Yu","year":"2017","unstructured":"Yu W, Yang K, Yao H, Sun X, Xu P (2017) Exploiting the complementary strengths of multi-layer CNN features for image retrieval. Neurocomputing 237:235\u2013241","journal-title":"Neurocomputing"},{"key":"3619_CR46","doi-asserted-by":"crossref","unstructured":"Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. Proc IEEE Conf Comput Vis Pattern Recognit:7132\u20137141","DOI":"10.1109\/CVPR.2018.00745"},{"key":"3619_CR47","unstructured":"Wang Q, Wu B, Zhu P, Li P, Zuo W, Hu Q (2020) Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR):11534\u201311542"},{"key":"3619_CR48","first-page":"740","volume-title":"Microsoft coco: common objects in context","author":"TY Lin","year":"2014","unstructured":"Lin TY, Maire M, Belongie S et al (2014) Microsoft coco: common objects in context. European conference on computer vision. Springer, Cham, pp 740\u2013755"},{"issue":"3","key":"3619_CR49","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M, Berg AC, Fei-Fei L (2015) Imagenet large scale visual recognition challenge. Int J Comput Vis 115(3):211\u2013252","journal-title":"Int J Comput Vis"},{"key":"3619_CR50","doi-asserted-by":"crossref","unstructured":"Real E, Shlens J, Mazzocchi S, Pan X, Vanhoucke V (2017). Youtube-boundingboxes: A large high-precision human-annotated data set for object detection in video. In Proc IEEE Conf Comput Vis Pattern Recognit pp 5296\u20135305","DOI":"10.1109\/CVPR.2017.789"},{"key":"3619_CR51","first-page":"205","volume-title":"Know your surroundings: exploiting scene information for object tracking","author":"G Bhat","year":"2020","unstructured":"Bhat G, Danelljan M, Gool LV et al (2020) Know your surroundings: exploiting scene information for object tracking. European Conference on Computer Vision. Springer, Cham, pp 205\u2013221"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-022-03619-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-022-03619-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-022-03619-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,11]],"date-time":"2023-01-11T11:14:57Z","timestamp":1673435697000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-022-03619-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,5,31]]},"references-count":51,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2023,2]]}},"alternative-id":["3619"],"URL":"https:\/\/doi.org\/10.1007\/s10489-022-03619-9","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"type":"print","value":"0924-669X"},{"type":"electronic","value":"1573-7497"}],"subject":[],"published":{"date-parts":[[2022,5,31]]},"assertion":[{"value":"10 April 2022","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 May 2022","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"All authors declare that No conflict of interest exists.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interests"}}]}}