{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T07:29:09Z","timestamp":1740122949163,"version":"3.37.3"},"reference-count":74,"publisher":"Springer Science and Business Media LLC","issue":"23","license":[{"start":{"date-parts":[[2024,1,10]],"date-time":"2024-01-10T00:00:00Z","timestamp":1704844800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,10]],"date-time":"2024-01-10T00:00:00Z","timestamp":1704844800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62261026"],"award-info":[{"award-number":["62261026"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62061019"],"award-info":[{"award-number":["62061019"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"the General Project of Jiangxi Natural Science Foundation","award":["20202BABL202014"],"award-info":[{"award-number":["20202BABL202014"]}]},{"DOI":"10.13039\/501100004479","name":"Natural Science Foundation of Jiangxi Province","doi-asserted-by":"publisher","award":["20212BAB202013"],"award-info":[{"award-number":["20212BAB202013"]}],"id":[{"id":"10.13039\/501100004479","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Key Project of Jiangxi Education Department","award":["GJJ201107"],"award-info":[{"award-number":["GJJ201107"]}]},{"DOI":"10.13039\/501100011343","name":"Key Laboratory of System Control and Information Processing","doi-asserted-by":"publisher","award":["Scip202106"],"award-info":[{"award-number":["Scip202106"]}],"id":[{"id":"10.13039\/501100011343","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-023-15256-6","type":"journal-article","created":{"date-parts":[[2024,1,10]],"date-time":"2024-01-10T06:02:36Z","timestamp":1704866556000},"page":"64015-64036","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["STransLOT: splitting-refusion transformer for low-light object tracking"],"prefix":"10.1007","volume":"83","author":[{"given":"Zhongwang","family":"Cai","sequence":"first","affiliation":[]},{"given":"Dunyun","family":"He","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0785-9382","authenticated-orcid":false,"given":"Zhen","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Fan","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Zhijian","family":"Yin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,1,10]]},"reference":[{"doi-asserted-by":"crossref","unstructured":"Bertinetto L, Valmadre J, Henriques JF, Vedaldi A, Torr PH (2016) Fully-convolutional siamese networks for object tracking. In: European conference on computer vision (pp 850-865). Springer, Cham","key":"15256_CR1","DOI":"10.1007\/978-3-319-48881-3_56"},{"doi-asserted-by":"crossref","unstructured":"Bhat G, Danelljan M, Gool LV, Timofte R (2019) Learning discriminative model prediction for tracking. In: ICCV","key":"15256_CR2","DOI":"10.1109\/ICCV.2019.00628"},{"unstructured":"Bhat G, Danelljan M, Gool LV, Timofte R (2019) Learning discriminative model prediction for tracking. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV). 1, 2, 3, 6, 7, 8, 13, 15, 16, 17","key":"15256_CR3"},{"doi-asserted-by":"crossref","unstructured":"Bhat G, Danelljan M, Van Gool L et al (2020) Know your surroundings: exploiting scene information for object tracking[C]. In: European conference on computer vision. Springer, Cham, pp 205\u2013221","key":"15256_CR4","DOI":"10.1007\/978-3-030-58592-1_13"},{"doi-asserted-by":"crossref","unstructured":"Bhat G, Johnander J, Danelljan M et al (2018) Unveiling the power of deep tracking[C]. In: Proceedings of the European Conference on Computer Vision (ECCV), pp 483\u2013498","key":"15256_CR5","DOI":"10.1007\/978-3-030-01216-8_30"},{"doi-asserted-by":"crossref","unstructured":"Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S (2020) End-to-end object detection with transformers. In: ECCV. 1, 2, 4, 5, 7","key":"15256_CR6","DOI":"10.1007\/978-3-030-58452-8_13"},{"doi-asserted-by":"crossref","unstructured":"Chen X, Yan B, Zhu J et al (2021) Transformer tracking[C]. In: Proceedings of the IEEE\/CVF Conference on computer vision and pattern recognition, pp 8126\u20138135","key":"15256_CR7","DOI":"10.1109\/CVPR46437.2021.00803"},{"doi-asserted-by":"crossref","unstructured":"Chen Z, Zhong B, Li G et al (2020) Siamese box adaptive network for visual tracking[C]. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 6668\u20136677","key":"15256_CR8","DOI":"10.1109\/CVPR42600.2020.00670"},{"unstructured":"Danelljan M, Bhat G (2019) PyTracking: visual tracking library based on PyTorch. https:\/\/github.com\/visionml\/pytracking. Accessed: Aug 01 2020. 2, 3, 7, 8, 13, 15, 16, 17","key":"15256_CR9"},{"doi-asserted-by":"crossref","unstructured":"Danelljan M, Bhat G, Khan FS, Felsberg M (2019) Atom: accurate tracking by overlap maximization. In: Proceedings of the IEEE\/CVF Conference on computer vision and pattern recognition, pp 4660\u20134669","key":"15256_CR10","DOI":"10.1109\/CVPR.2019.00479"},{"doi-asserted-by":"crossref","unstructured":"Danelljan M, Bhat G, Shahbaz Khan F, Felsberg M (2017) Eco: efficient convolution operators for tracking. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6638\u20136646","key":"15256_CR11","DOI":"10.1109\/CVPR.2017.733"},{"doi-asserted-by":"crossref","unstructured":"Danelljan M, Gool LV, Timofte R (2020) Probabilistic regression for visual tracking. 2, 3, 7, 8, 13, 15, 16, 17","key":"15256_CR12","DOI":"10.1109\/CVPR42600.2020.00721"},{"doi-asserted-by":"crossref","unstructured":"Dong XP, Shen JB (2018) Triplet loss in siamese network for object tracking. In: ECCV, p 2","key":"15256_CR13","DOI":"10.1007\/978-3-030-01261-8_28"},{"doi-asserted-by":"crossref","unstructured":"Dong X, Shen J, Shao L et al (2020) CLNet: a compact latent network for fast adjusting Siamese trackers[C]. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XX. Cham: Springer International Publishing, pp 378\u2013395","key":"15256_CR14","DOI":"10.1007\/978-3-030-58565-5_23"},{"issue":"99","key":"15256_CR15","first-page":"1","volume":"PP","author":"X Dong","year":"2019","unstructured":"Dong X, Shen J, Wang W et al (2019) Dynamical hyperparameter optimization via deep reinforcement learning in tracking[J]. IEEE Trans Pattern Anal Mach Intell PP(99):1\u20131","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"7","key":"15256_CR16","doi-asserted-by":"publisher","first-page":"3516","DOI":"10.1109\/TIP.2019.2898567","volume":"28","author":"X Dong","year":"2019","unstructured":"Dong X, Shen J, Wu D et al (2019) Quadruplet network with one-shot learning for fast visual object tracking[J]. IEEE Trans Image Process 28 (7):3516\u20133527","journal-title":"IEEE Trans Image Process"},{"doi-asserted-by":"crossref","unstructured":"Fan H, Lin L, Yang F et al (2019) Lasot: A high-quality benchmark for large-scale single object tracking[C]. In: Proceedings of the IEEE\/CVF Conference on computer vision and pattern recognition, pp 5374\u20135383","key":"15256_CR17","DOI":"10.1109\/CVPR.2019.00552"},{"unstructured":"Glorot X, Bengio Y (2010) Understanding the difficulty of training deep feedforward neural networks. In: ICAIS, p 6","key":"15256_CR18"},{"unstructured":"G\u00fcndo\u011fdu E, Alatan AA (2016) The Visual Object Tracking VOT2016 challenge results[J]","key":"15256_CR19"},{"doi-asserted-by":"crossref","unstructured":"Guo Q, Feng W, Zhou C, Huang R, Wan L, Wang S (2017) Learning dynamic siamese network for visual object tracking. In: ICCV, vol 2, p 7","key":"15256_CR20","DOI":"10.1109\/ICCV.2017.196"},{"doi-asserted-by":"crossref","unstructured":"Guo D, Wang J, Cui Y et al (2020) SiamCAR: Siamese fully convolutional classification and regression for visual tracking[C]. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 6269\u20136277","key":"15256_CR21","DOI":"10.1109\/CVPR42600.2020.00630"},{"doi-asserted-by":"crossref","unstructured":"Han W, Dong X, Khan FS et al (2021) Learning to fuse asymmetric feature maps in siamese trackers[C]. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 16570\u201316580","key":"15256_CR22","DOI":"10.1109\/CVPR46437.2021.01630"},{"doi-asserted-by":"crossref","unstructured":"He AF, Luo C, Tian XM, Zeng WJ (2018) A twofold siamese network for real-time object tracking. In: CVPR, vol 1, p 2","key":"15256_CR23","DOI":"10.1109\/CVPR.2018.00508"},{"doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S et al (2016) Deep residual learning for image recognition[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778","key":"15256_CR24","DOI":"10.1109\/CVPR.2016.90"},{"issue":"3","key":"15256_CR25","doi-asserted-by":"publisher","first-page":"583","DOI":"10.1109\/TPAMI.2014.2345390","volume":"37","author":"JF Henriques","year":"2014","unstructured":"Henriques JF, Caseiro R, Martins P, Batista J (2014) High-speed tracking with kernelized correlation filters. IEEE Trans Pattern Anal Mach Intell 37(3):583\u2013596","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"unstructured":"Huang L, Zhao X, Huang K (2019) Got-10k: a large high-diversity benchmark for generic object tracking in the wild[J]. IEEE Trans Pattern Anal Mach Intell","key":"15256_CR26"},{"doi-asserted-by":"crossref","unstructured":"Jung I, Son J et al (2018) Real-time MDNet. In: ECCV, pp 83\u201398","key":"15256_CR27","DOI":"10.1007\/978-3-030-01225-0_6"},{"doi-asserted-by":"crossref","unstructured":"Kiani Galoogahi H, Fagg A, Huang C et al (2017) Need for speed: a benchmark for higher frame rate object tracking[C]. In: Proceedings of the IEEE International conference on computer vision, pp 1125\u20131134","key":"15256_CR28","DOI":"10.1109\/ICCV.2017.128"},{"unstructured":"Kristan M, Leonardis A, Matas J et al (2018) The sixth visual object tracking vot2018 challenge results[C]. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops, pp 0\u20130","key":"15256_CR29"},{"doi-asserted-by":"crossref","unstructured":"Kristan M, Matas J, Leonardis A et al (2019) The seventh visual object tracking vot2019 challenge results[C]. In: Proceedings of the IEEE\/CVF International conference on computer vision workshops, pp 0\u20130","key":"15256_CR30","DOI":"10.1109\/ICCVW.2019.00276"},{"issue":"9","key":"15256_CR31","doi-asserted-by":"publisher","first-page":"1429","DOI":"10.1109\/TMM.2015.2455418","volume":"17","author":"KH Lee","year":"2015","unstructured":"Lee KH, Hwang JN (2015) On-road pedestrian tracking across multiple driving recorders. IEEE Trans Multimed 17(9):1429\u20131438","journal-title":"IEEE Trans Multimed"},{"doi-asserted-by":"crossref","unstructured":"Li B, Wu W, Wang Q, Zhang F, Xing J, Yan J (2019) Siamrpn++: evolution of siamese visual tracking with very deep networks. In: Proceedings of the IEEE\/CVF Conference on computer vision and pattern recognition, pp 4282\u20134291","key":"15256_CR32","DOI":"10.1109\/CVPR.2019.00441"},{"doi-asserted-by":"crossref","unstructured":"Li B, Yan J, Wu W, Zhu Z, Hu X (2018) High performance visual tracking with siamese region proposal network. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8971\u20138980","key":"15256_CR33","DOI":"10.1109\/CVPR.2018.00935"},{"doi-asserted-by":"crossref","unstructured":"Liang Z, Shen J (2019) Local semantic siamese networks for fast tracking[J]. IEEE Trans Image Process, pp 29(99)","key":"15256_CR34","DOI":"10.1109\/TIP.2019.2959256"},{"doi-asserted-by":"crossref","unstructured":"Liang C, Wang W, Zhou T, Miao J, Luo Y, Yang Y (2022) Local-global context aware transformer for language-guided video segmentation. arXiv:2203.09773","key":"15256_CR35","DOI":"10.1109\/TPAMI.2023.3262578"},{"doi-asserted-by":"crossref","unstructured":"Lin TY, Maire M, Belongie S et al (2014) Microsoft coco: common objects in context[C]. In: European conference on computer vision. Springer, Cham, pp 740\u2013755","key":"15256_CR36","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"15256_CR37","doi-asserted-by":"publisher","first-page":"650","DOI":"10.1016\/j.patcog.2016.06.008","volume":"61","author":"KG Lore","year":"2017","unstructured":"Lore KG, Akintayo A, Sarkar S (2017) LLNEt: a deep autoencoder approach to natural low-light image enhancement. Pattern Recogn 61:650\u2013662","journal-title":"Pattern Recogn"},{"issue":"99","key":"15256_CR38","first-page":"1","volume":"PP","author":"X Lu","year":"2020","unstructured":"Lu X, Ma C, Shen J et al (2020) Deep object tracking with shrinkage loss[J]. IEEE Trans Pattern Anal Mach Intell PP(99):1\u20131","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"doi-asserted-by":"crossref","unstructured":"Luscher C, Beck E, Irie K, Kitza M, Michel W, Zeyer A, Schluter R, Ney H (2019) RWTH ASR Systems For librispeech: hybrid vs attention. In: INTERSPEECH, p 2","key":"15256_CR39","DOI":"10.21437\/Interspeech.2019-1780"},{"issue":"23","key":"15256_CR40","doi-asserted-by":"publisher","first-page":"231104","DOI":"10.1063\/1.4809836","volume":"10;102","author":"OS Magana-Loaiza","year":"2013","unstructured":"Magana-Loaiza OS, Howland GA, Malik M, Howell JC, Boyd RW (2013) Compressive object tracking using entangled photons. Appl Phys Lett 10;102(23):231104","journal-title":"Appl Phys Lett"},{"doi-asserted-by":"crossref","unstructured":"Mayer C, Danelljan M, Bhat G et al (2022) Transforming model prediction for tracking[C]. In: Proceedings of the IEEE\/CVF Conference on computer vision and pattern recognition, pp 8731\u20138740","key":"15256_CR41","DOI":"10.1109\/CVPR52688.2022.00853"},{"doi-asserted-by":"crossref","unstructured":"Mayer C, Danelljan M, Paudel DP et al (2021) Learning target candidate association to keep track of what not to track[J]. arXiv:2103.16556","key":"15256_CR42","DOI":"10.1109\/ICCV48922.2021.01319"},{"doi-asserted-by":"crossref","unstructured":"Meinhardt T, Kirillov A, Leal-Taixe L, Feichtenhofer C (2021) TrackFormer: multi-object tracking with transformers. arXiv:2101.02702. 1, 2, 7, 8","key":"15256_CR43","DOI":"10.1109\/CVPR52688.2022.00864"},{"doi-asserted-by":"crossref","unstructured":"Mueller M, Smith N, Ghanem B (2016) A benchmark and simulator for uav tracking[C]. In: European conference on computer vision. Springer, Cham, pp 445\u2013461","key":"15256_CR44","DOI":"10.1007\/978-3-319-46448-0_27"},{"doi-asserted-by":"crossref","unstructured":"Muller M, Bibi A, Giancola S et al (2018) Trackingnet: a large-scale dataset and benchmark for object tracking in the wild[C]. In: Proceedings of the European Conference on Computer Vision (ECCV), pp 300\u2013317","key":"15256_CR45","DOI":"10.1007\/978-3-030-01246-5_19"},{"doi-asserted-by":"crossref","unstructured":"Nam H, Han B (2016) Learning multi-domain convolutional neural networks for visual tracking. In: CVPR","key":"15256_CR46","DOI":"10.1109\/CVPR.2016.465"},{"unstructured":"Parmar N, Vaswani A, Uszkoreit J, Kaiser L, Shazeer N, Ku A, Tran D (2018) Image transformer. In: ICML, p 2","key":"15256_CR47"},{"unstructured":"Radford A, Wu J, Child R, Luan D, Amodei D, Sutskever I (2019) Language models are unsupervised multitask learners. OpenAI blog","key":"15256_CR48"},{"key":"15256_CR49","first-page":"91","volume":"28","author":"S Ren","year":"2015","unstructured":"Ren S, He K, Girshick R et al (2015) Faster r-cnn: towards real-time object detection with region proposal networks[J]. Adv Neural Inf Process Syst 28:91\u201399","journal-title":"Adv Neural Inf Process Syst"},{"issue":"12","key":"15256_CR50","doi-asserted-by":"publisher","first-page":"8896","DOI":"10.1109\/TPAMI.2021.3127492","volume":"44","author":"J Shen","year":"2021","unstructured":"Shen J, Liu Y, Dong X et al (2021) Distilled Siamese networks for visual tracking[J]. IEEE Trans Pattern Anal Mach Intell 44(12):8896\u20138909","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"7","key":"15256_CR51","doi-asserted-by":"publisher","first-page":"3068","DOI":"10.1109\/TCYB.2019.2936503","volume":"50","author":"J Shen","year":"2019","unstructured":"Shen J, Tang X, Dong X et al (2019) Visual object tracking by hierarchical attention siamese network[J]. IEEE Trans Cybern 50(7):3068\u20133080","journal-title":"IEEE Trans Cybern"},{"unstructured":"Sun P, Jiang Y, Zhang R, Xie E, Cao J, Hu X, Kong T, Yuan Z, Wang C, Luo P (2020) TransTrack:, Multiple-object tracking with transformer. arXiv:2012.15460","key":"15256_CR52"},{"unstructured":"Synnaeve G, Xu Q, Kahn J, Grave E, Likhomanenko T, Pratap V, Sriram A, Liptchinsky V, Collobert R (2019) End-to-end ASR: from supervised to semi-supervised learning with modern architectures. CoRR, arXiv:1911.08460","key":"15256_CR53"},{"doi-asserted-by":"crossref","unstructured":"Valmadre J, Bertinetto L, Henriques JF, Vedaldi A, Torr PH (2017) End-to-end representation learning for correlation filter based tracking. In: CVPR. 1, 2, 6, 7","key":"15256_CR54","DOI":"10.1109\/CVPR.2017.531"},{"unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Polosukhin I (2017) Attention is all you need. In: Advances in neural information processing systems, pp 5998\u20136008","key":"15256_CR55"},{"doi-asserted-by":"crossref","unstructured":"Voigtlaender P, Luiten J, Torr PHS et al (2020) Siam r-cnn: Visual tracking by re-detection[C]. In: Proceedings of the IEEE\/CVF Conference on computer vision and pattern recognition, pp 6578\u20136588","key":"15256_CR56","DOI":"10.1109\/CVPR42600.2020.00661"},{"key":"15256_CR57","doi-asserted-by":"publisher","first-page":"108498","DOI":"10.1016\/j.patcog.2021.108498","volume":"1;124","author":"C Wang","year":"2022","unstructured":"Wang C, Wang X, Zhang J, Zhang L, Bai X, Ning X, Zhou J, Hancock E (2022) Uncertainty estimation for stereo matching based on evidential deep learning. Pattern Recognit 1;124:108498","journal-title":"Pattern Recognit"},{"doi-asserted-by":"crossref","unstructured":"Wang Q, Zhang L, Bertinetto L et al (2019) Fast online object tracking and segmentation: a unifying approach[C]. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 1328\u20131338","key":"15256_CR58","DOI":"10.1109\/CVPR.2019.00142"},{"doi-asserted-by":"crossref","unstructured":"Wang S, Zhou T, Lu Y, Di H (2022) Detail preserving transformer for light field image super-resolution. In: Proceedings AAAI Conference Artificial Intelligent","key":"15256_CR59","DOI":"10.1609\/aaai.v36i3.20153"},{"doi-asserted-by":"crossref","unstructured":"Wang N, Zhou W, Wang J et al (2021) Transformer meets tracker: exploiting temporal context for robust visual tracking[C]. In: Proceedings of the IEEE\/CVF Conference on computer vision and pattern recognition, pp 1571\u20131580","key":"15256_CR60","DOI":"10.1109\/CVPR46437.2021.00162"},{"doi-asserted-by":"crossref","unstructured":"Wang H, Zhu Y, Adam H, Yuille A, Chen L-C (2020) MaX-DeepLab: end-to-end panoptic segmentation with mask transformers. arXiv:2012.00759.2","key":"15256_CR61","DOI":"10.1109\/CVPR46437.2021.00542"},{"issue":"9","key":"15256_CR62","doi-asserted-by":"publisher","first-page":"1834","DOI":"10.1109\/TPAMI.2014.2388226","volume":"37","author":"Y Wu","year":"2015","unstructured":"Wu Y, Lim J, Yang MH (2015) Object tracking benchmark[J]. IEEE Trans Pattern Anal Mach Intell 37(9):1834\u20131848","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"doi-asserted-by":"crossref","unstructured":"Xie F, Wang C, Wang G et al (2022) Correlation-aware deep tracking[C]. In: Proceedings of the IEEE\/CVF Conference on computer vision and pattern recognition, pp 8751\u20138760","key":"15256_CR63","DOI":"10.1109\/CVPR52688.2022.00855"},{"doi-asserted-by":"crossref","unstructured":"Xing J, Ai H, Lao S (2010) Multiple human tracking based on multi-view upper-body detection and discriminative learning. In: 2010 20th International conference on pattern recognition (pp 1698\u20131701). IEEE","key":"15256_CR64","DOI":"10.1109\/ICPR.2010.420"},{"doi-asserted-by":"crossref","unstructured":"Xu Y, Wang Z, Li Z et al (2020) SiamFC++: towards robust and accurate visual tracking with target estimation guidelines[C]. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 34, pp 12549\u201312556","key":"15256_CR65","DOI":"10.1609\/aaai.v34i07.6944"},{"doi-asserted-by":"crossref","unstructured":"Yan B, Peng H, Fu J et al (2021) Learning spatio-temporal transformer for visual tracking[J]. arXiv:2103.17154","key":"15256_CR66","DOI":"10.1109\/ICCV48922.2021.01028"},{"doi-asserted-by":"crossref","unstructured":"Yang T, Chan AB (2018) Learning dynamic memory networks for object tracking. In: ECCV","key":"15256_CR67","DOI":"10.1007\/978-3-030-01240-3_10"},{"doi-asserted-by":"crossref","unstructured":"Zhang L, Gonzalez-Garcia A, Weijer JVD, Danelljan M, Shahbaz Khan F (2019) Learning the model update for siamese trackers. In: ICCV","key":"15256_CR68","DOI":"10.1109\/ICCV.2019.00411"},{"issue":"1","key":"15256_CR69","first-page":"1","volume":"25;11","author":"J Zhang","year":"2021","unstructured":"Zhang J, Huang B, Ye Z, Kuang LD, Ning X (2021) Siamese anchor-free object tracking with multiscale spatial attentions. Sci Rep 25;11(1):1\u20134","journal-title":"Sci Rep"},{"doi-asserted-by":"crossref","unstructured":"Zhang Z, Peng H (2019) Deeper and wider siamese networks for real-time visual tracking[C]. In: Proceedings of the IEEE\/CVF Conference on computer vision and pattern recognition, pp 4591\u20134600","key":"15256_CR70","DOI":"10.1109\/CVPR.2019.00472"},{"doi-asserted-by":"crossref","unstructured":"Zhang Z, Peng H, Fu J et al (2020) Ocean: object-aware anchor-free tracking[C]. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XXI 16. Springer International Publishing, pp 771\u2013787","key":"15256_CR71","DOI":"10.1007\/978-3-030-58589-1_46"},{"doi-asserted-by":"crossref","unstructured":"Zhang G, Vela PA (2015) Good features to track for visual slam. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1373\u20131382","key":"15256_CR72","DOI":"10.1109\/CVPR.2015.7298743"},{"unstructured":"Zhu X, Su W, Lu L et al (2020) Deformable detr: deformable transformers for end-to-end object detection[J]. arXiv:2010.04159","key":"15256_CR73"},{"doi-asserted-by":"crossref","unstructured":"Zhu Z, Wang Q, Li B et al (2018) Distractor-aware siamese networks for visual object tracking[C]. In: Proceedings of the European Conference on Computer Vision (ECCV), pp 101\u2013117","key":"15256_CR74","DOI":"10.1007\/978-3-030-01240-3_7"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-15256-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-023-15256-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-15256-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,5]],"date-time":"2024-07-05T16:18:11Z","timestamp":1720196291000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-023-15256-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1,10]]},"references-count":74,"journal-issue":{"issue":"23","published-online":{"date-parts":[[2024,7]]}},"alternative-id":["15256"],"URL":"https:\/\/doi.org\/10.1007\/s11042-023-15256-6","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2024,1,10]]},"assertion":[{"value":"11 January 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 March 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 April 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 January 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article. The datasets generated during and\/or analysed during the current study are available from the corresponding author on reasonable request.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"<!--Emphasis Type='Bold' removed-->Conflict of interests"}}]}}