{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,31]],"date-time":"2025-12-31T00:37:14Z","timestamp":1767141434209,"version":"build-2238731810"},"reference-count":65,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2022,8,13]],"date-time":"2022-08-13T00:00:00Z","timestamp":1660348800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,8,13]],"date-time":"2022-08-13T00:00:00Z","timestamp":1660348800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100005153","name":"China National Funds for Distinguished Young Scientists","doi-asserted-by":"publisher","award":["11664005"],"award-info":[{"award-number":["11664005"]}],"id":[{"id":"10.13039\/501100005153","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100010899","name":"Joint Fund of Coal","doi-asserted-by":"publisher","award":["2021YFF0500901"],"award-info":[{"award-number":["2021YFF0500901"]}],"id":[{"id":"10.13039\/501100010899","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Process Lett"],"published-print":{"date-parts":[[2023,4]]},"DOI":"10.1007\/s11063-022-10951-1","type":"journal-article","created":{"date-parts":[[2022,8,13]],"date-time":"2022-08-13T08:06:34Z","timestamp":1660377994000},"page":"1555-1580","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["A Lightweight Neural Learning Algorithm for Real-Time Facial Feature Tracking System via Split-Attention and Heterogeneous Convolution"],"prefix":"10.1007","volume":"55","author":[{"given":"Yuandong","family":"Ma","sequence":"first","affiliation":[]},{"given":"Qing","family":"Song","sequence":"additional","affiliation":[]},{"given":"Mengjie","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Xiaotong","family":"Zhu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,8,13]]},"reference":[{"key":"10951_CR1","doi-asserted-by":"crossref","unstructured":"S, Ayd\u0131n.: Deep Learning Classification of Neuro-Emotional Phase Domain Complexity Levels Induced by Affective Video Film Clips.IEEE Journal of Biomedical and Health Informatics.24 (6), pp.1695\u20131702","DOI":"10.1109\/JBHI.2019.2959843"},{"key":"10951_CR2","doi-asserted-by":"publisher","unstructured":"S, Ayd\u0131n.: Cross-validated Adaboost Classification of Emotion Regulation Strategies Identified by Spectral Coherence in Resting-State. Neuroinform https:\/\/doi.org\/10.1007\/s12021-021-09542-7","DOI":"10.1007\/s12021-021-09542-7"},{"issue":"6","key":"10951_CR3","first-page":"390","volume":"37","author":"X Yu","year":"2020","unstructured":"Yu X, Li Y, Zhang S, Xue C, Yu W (2020) Estimation of human impedance and motion intention for constrained human\u2013robot interaction. Neurocomputing 37(6):390\u2013403","journal-title":"Neurocomputing"},{"key":"10951_CR4","first-page":"1","volume-title":"ACM SIGGRAPH 2011 papers (SIGGRAPH \u201811)","author":"SLG Eduardo","year":"2011","unstructured":"Eduardo SLG, Manuel MO (2011) Domain transform for edge-aware image and video processing. ACM SIGGRAPH 2011 papers (SIGGRAPH \u201811). Association for Computing Machinery, New York, NY, USA, pp 1\u201312"},{"issue":"3","key":"10951_CR5","first-page":"1","volume":"20","author":"J Wu","year":"2020","unstructured":"Wu J, Ji Y, Sun X et al (2020) Price Regulation Mechanism of Travelers\u2019 Travel Mode Choice in the Driverless Transportation Network. J Adv Transp 20(3):1\u20139","journal-title":"J Adv Transp"},{"key":"10951_CR6","doi-asserted-by":"publisher","first-page":"701","DOI":"10.1007\/s00371-019-01652-3","volume":"36","author":"M Zolfaghari","year":"2020","unstructured":"Zolfaghari M, Ghanei-Yakhdan H, Yazdi M (2020) Real-time object tracking based on an adaptive transition model and extended Kalman filter to handle full occlusion. Vis Comput 36:701\u2013715","journal-title":"Vis Comput"},{"key":"10951_CR7","doi-asserted-by":"publisher","first-page":"887","DOI":"10.1007\/s11063-016-9554-6","volume":"45","author":"MN Islam","year":"2017","unstructured":"Islam MN, Loo CK, Seera M (2017) Incremental Clustering-Based Facial Feature Tracking Using Bayesian ART. Neural Process Lett 45:887\u2013911","journal-title":"Neural Process Lett"},{"issue":"8","key":"10951_CR8","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TPAMI.2018.2877675","volume":"41","author":"L Sheng","year":"2019","unstructured":"Sheng L, Cai J, Cham T et al (2019) Visibility Constrained Generative Model for Depth-Based 3D Facial Pose Tracking. IEEE Trans Pattern Anal Mach Intell 41(8):2614\u20132623","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"10951_CR9","doi-asserted-by":"publisher","first-page":"133","DOI":"10.1023\/A:1009623602553","volume":"7","author":"G Marcone","year":"1998","unstructured":"Marcone G, Martinelli G, Lancetti L (1998) Eye Tracking in Image Sequences by Competitive Neural Networks. Neural Process Lett 7:133\u2013138","journal-title":"Neural Process Lett"},{"key":"10951_CR10","doi-asserted-by":"publisher","first-page":"881","DOI":"10.1007\/s00371-020-01839-z","volume":"37","author":"W Zhang","year":"2021","unstructured":"Zhang W, Du Y, Chen Z et al (2021) Robust adaptive learning with siamese network architecture for visual tracking. Vis Comput 37:881\u2013894","journal-title":"Vis Comput"},{"issue":"12","key":"10951_CR12","first-page":"749","volume":"32","author":"Z Ren","year":"2018","unstructured":"Ren Z, Li J, Zhang X et al (2018) A Face Tracking Method in Videos Based on Convolutional Neural. Networks Int J Pattern Recognit Artif Intell 32(12):749\u2013762","journal-title":"Networks Int J Pattern Recognit Artif Intell"},{"key":"10951_CR13","doi-asserted-by":"crossref","unstructured":"Danelljan M, Robinson A, Khan FS et al (2016) : Beyond correlation filters: learning continuous convolution operators for visual tracking. In: European conference on computer vision (ECCV), pp.\u00a0472\u2013488.Springer, Cham","DOI":"10.1007\/978-3-319-46454-1_29"},{"key":"10951_CR14","doi-asserted-by":"crossref","unstructured":"Danelljan M, Bhat G, Shahbaz KF et al (2017) : ECO: efficient convolution operators for tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a06638\u20136646","DOI":"10.1109\/CVPR.2017.733"},{"key":"10951_CR15","doi-asserted-by":"crossref","unstructured":"Luca B, Jack V, Jo\u02dcao FH, Andrea V, Philip HST (2016) : Fully-convolutional siamese networks for object tracking. In: European conference on computer vision (ECCV), pp.\u00a0850\u2013865. Springer, Cham","DOI":"10.1007\/978-3-319-48881-3_56"},{"key":"10951_CR16","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1016\/j.image.2019.02.009","volume":"74","author":"OJ Makhura","year":"2019","unstructured":"Makhura OJ, Woods JC (2019) Learn-select-track: an approach to multi-object tracking. Signal Process Image Commun 74:153\u2013161","journal-title":"Signal Process Image Commun"},{"issue":"14","key":"10951_CR17","doi-asserted-by":"publisher","first-page":"142","DOI":"10.1016\/j.neucom.2019.01.116","volume":"393","author":"L Lin","year":"2020","unstructured":"Lin L, Liu B, Xiao Y (2020) COB method with online learning for object tracking. Neurocomputing 393(14):142\u2013155","journal-title":"Neurocomputing"},{"issue":"Apra","key":"10951_CR18","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1016\/j.jvcir.2019.01.032","volume":"60","author":"B Tu","year":"2019","unstructured":"Tu B, Kuang W, Shang Y et al (2019) A multi-view object tracking using triplet model. J visual communication image representation 60(Apra):64\u201368","journal-title":"J visual communication image representation"},{"key":"10951_CR19","doi-asserted-by":"crossref","unstructured":"Li B, Yan J, Wu W, Zhu Z, Hu X (2018) : High performance visual tracking with siamese region proposal network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.8971\u20138980","DOI":"10.1109\/CVPR.2018.00935"},{"key":"10951_CR20","doi-asserted-by":"crossref","unstructured":"Li B, Wu W, Wang Q, Zhang F, Xing J, Yan J (2019) : SiamRPN++: Evolution of siamese visual tracking with very deep networks. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.1\u20137","DOI":"10.1109\/CVPR.2019.00441"},{"key":"10951_CR21","doi-asserted-by":"crossref","unstructured":"Xu Y, Wang Z, Li Z, Yu, Gang (2020) : SiamFC++: towards robust and accurate visual tracking with target estimation guidelines. In: Association for the Advance of Artificial Intelligence (AAAI), pp.256\u2013267","DOI":"10.1609\/aaai.v34i07.6944"},{"key":"10951_CR22","doi-asserted-by":"crossref","unstructured":"Guo D, Wang J, Cui Y, Wang Z, Chen S (2020) : SiamCAR: Siamese fully convolutional classification and regression for visual tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a06268\u20136276","DOI":"10.1109\/CVPR42600.2020.00630"},{"key":"10951_CR23","doi-asserted-by":"crossref","unstructured":"Held D, Thrun S, Savarese S (2016) : Learning to track at 100 fps with deep regression networks. In: European conference on computer vision (ECCV), pp.\u00a0749\u2013765. Springer, Cham","DOI":"10.1007\/978-3-319-46448-0_45"},{"key":"10951_CR24","doi-asserted-by":"crossref","unstructured":"Wei S, Ren J (2016) : Real-time tracking of non-rigid objects. In: International Conference on Communication and Information Systems pp.11\u201315","DOI":"10.1145\/3023924.3023944"},{"key":"10951_CR25","doi-asserted-by":"publisher","first-page":"745","DOI":"10.1007\/s11063-016-9509-y","volume":"45","author":"WL Zheng","year":"2017","unstructured":"Zheng WL, Shen SC, Lu BL (2017) Online Depth Image-Based Object Tracking with Sparse Representation and Object Detection. Neural Process Lett 45:745\u2013758","journal-title":"Neural Process Lett"},{"issue":"23","key":"10951_CR26","doi-asserted-by":"publisher","first-page":"16751","DOI":"10.1007\/s11042-020-08889-4","volume":"79","author":"Z Lian","year":"2020","unstructured":"Lian Z, Shao S, Huang C (2020) A real-time face tracking system based on multiple information fusion. Multimedia Tools and Applications 79(23):16751\u201316769","journal-title":"Multimedia Tools and Applications"},{"key":"10951_CR27","doi-asserted-by":"crossref","unstructured":"Osuna E, Freund R, Girosit F (1997) : Training support vector machines: an application to face detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a0130\u2013136","DOI":"10.1109\/CVPR.1997.609310"},{"issue":"8","key":"10951_CR28","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3469029","volume":"54","author":"MG Sarwar Murshed","year":"2022","unstructured":"Sarwar Murshed MG, Murphy C, Hou DQ, Khan N, Ganesh A, Hussain F (2022) Machine Learning at the Network Edge: A Survey. ACM Comput Surv 54(8):1\u201337","journal-title":"ACM Comput Surv"},{"key":"10951_CR29","doi-asserted-by":"crossref","unstructured":"Chen J, Ran X (2019) : Deep Learning With Edge Computing: A Review. in Proceedings of the IEEE, 107 (8), 1655\u20131674","DOI":"10.1109\/JPROC.2019.2921977"},{"key":"10951_CR30","doi-asserted-by":"crossref","unstructured":"Zhou Z, Chen X, Li E, Zeng L, Luo K, Zhang J (2019) : Edge Intelligence: Paving the Last Mile of Artificial Intelligence With Edge Computing. in Proceedings of the IEEE, 107 (8), 1738\u20131762","DOI":"10.1109\/JPROC.2019.2918951"},{"key":"10951_CR31","doi-asserted-by":"crossref","unstructured":"Zhang Z, Peng H (2019) : Deeper and wider siamese networks for real-time visual tracking. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.678\u2013685","DOI":"10.1109\/CVPR.2019.00472"},{"key":"10951_CR32","doi-asserted-by":"publisher","unstructured":"Singh et al (2019) : HetConv: Heterogeneous Kernel-Based Convolutions for Deep CNNs. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a04830\u20134839 (2019) doi: https:\/\/doi.org\/10.1109\/CVPR.2019.00497","DOI":"10.1109\/CVPR.2019.00497"},{"key":"10951_CR33","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) : Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a0770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"issue":"2","key":"10951_CR34","doi-asserted-by":"publisher","first-page":"114","DOI":"10.5391\/IJFIS.2017.17.2.114","volume":"17","author":"Y Hwang","year":"2017","unstructured":"Hwang Y, Jeong MH, Oh SR, Yoon C (2017) Adaptive mean shift based face tracking by coupled support map. Int J Fuzzy Logic Intell Syst 17(2):114\u2013120","journal-title":"Int J Fuzzy Logic Intell Syst"},{"key":"10951_CR35","doi-asserted-by":"publisher","first-page":"33","DOI":"10.1007\/s11063-012-9267-4","volume":"37","author":"M Mentzelopoulos","year":"2013","unstructured":"Mentzelopoulos M, Psarrou A, Angelopoulou A et al (2013) Active Foreground Region Extraction and Tracking for Sports Video Annotation. Neural Process Lett 37:33\u201346","journal-title":"Neural Process Lett"},{"issue":"1","key":"10951_CR36","doi-asserted-by":"publisher","first-page":"1247","DOI":"10.1007\/s11042-020-09780-y","volume":"80","author":"DY Huang","year":"2021","unstructured":"Huang DY, Chen CH, Chen TY, Hu WC, Guo ZB, Wen CK (2021) High-efficiency face detection and tracking method for numerous pedestrians through face candidate generation. Multimedia Tools and Applications 80(1):1247\u20131272","journal-title":"Multimedia Tools and Applications"},{"key":"10951_CR37","doi-asserted-by":"crossref","unstructured":"Tao R, Gavves E, Smeulders AWM (2016) : Siamese instance search for tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a01420\u20131429","DOI":"10.1109\/CVPR.2016.158"},{"key":"10951_CR38","doi-asserted-by":"crossref","unstructured":"Chen Z, Zhong B, Li G, Zhang S, Ji R (2020) : Siamese box adaptive network for visual tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.6667\u20136676","DOI":"10.1109\/CVPR42600.2020.00670"},{"key":"10951_CR39","unstructured":"Paul V, Jonathon L, Philip HST, Bastian L (2020) : Siam R-CNN: Visual tracking by re-detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.6577\u20136587"},{"key":"10951_CR40","doi-asserted-by":"crossref","unstructured":"Zhang Z, Peng H, Fu J, Li B, Hu W (2020) : Ocean: Object-aware anchor-free tracking. In: European conference on computer vision (ECCV), pp.771\u2013787","DOI":"10.1007\/978-3-030-58589-1_46"},{"key":"10951_CR41","unstructured":"Cai H, Gan C, Wang T, Zhang Z, Han S (2020) : Once-for-all: Train one network and specialize it for efficient deployment. In: International Conference on Learning Representations (ICLR),.751\u2013762"},{"key":"10951_CR42","doi-asserted-by":"crossref","unstructured":"Lu J, Xiong C, Parikh D, Socher R (2017) : Knowing when to look: adaptive attention via a visual sentinel for image captioning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.3242\u20133250","DOI":"10.1109\/CVPR.2017.345"},{"key":"10951_CR43","doi-asserted-by":"crossref","unstructured":"Chen L, Yang Y, Wang J, Xu W, Yuille AL (2016) : Attention to scale: scale-aware semantic image segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.3640\u20133649","DOI":"10.1109\/CVPR.2016.396"},{"issue":"11","key":"10951_CR44","doi-asserted-by":"publisher","first-page":"3475","DOI":"10.1109\/TITS.2017.2775628","volume":"19","author":"H Fan","year":"2018","unstructured":"Fan H, Mei X, Prokhorov D, Ling H (2018) Multi-level contextual RNNs with attention model for scene labeling. Intell Transp Syst IEEE Trans on 19(11):3475\u20133485","journal-title":"Intell Transp Syst IEEE Trans on"},{"key":"10951_CR45","doi-asserted-by":"crossref","unstructured":"Xu H, Saenko K (2016) : Ask, attend and answer: exploring question-guided spatial attention for visual question answering. In: European conference on computer vision (ECCV), pp.451\u2013466","DOI":"10.1007\/978-3-319-46478-7_28"},{"key":"10951_CR46","doi-asserted-by":"crossref","unstructured":"Choi J, Chang HJ, Yun S, Fischer T, Demiris Y, Choi JY (2017) : Attentional correlation filter network for adaptive visual tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.4828\u20134837","DOI":"10.1109\/CVPR.2017.513"},{"key":"10951_CR47","doi-asserted-by":"crossref","unstructured":"Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) : Going deeper with convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a01\u20139","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"10951_CR48","doi-asserted-by":"crossref","unstructured":"Xie S, Girshick R, Doll\u00b4 ar P, Tu Z, He K (2017) : Aggregated residual transformations for deep neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a01492\u20131500","DOI":"10.1109\/CVPR.2017.634"},{"issue":"8","key":"10951_CR49","doi-asserted-by":"publisher","first-page":"2011","DOI":"10.1109\/TPAMI.2019.2913372","volume":"42","author":"J Hu","year":"2020","unstructured":"Hu J, Shen L, Sun G (2020) Squeeze-and-excitation networks. IEEE Trans Pattern Anal Mach Intell 42(8):2011\u20132023","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"10951_CR50","doi-asserted-by":"crossref","unstructured":"Li X, Wang W, Hu X, Yang J (2019) : Selective kernel networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.1011\u20131022","DOI":"10.1109\/CVPR.2019.00060"},{"key":"10951_CR51","unstructured":"Han S, Mao H, Dally WJ (2016) : Deep compression: Compressing deep neural networks with pruning, trained quantization and huff man coding. In: International Conference on Learning Representations (ICLR), pp.\u00a0458\u2013467"},{"key":"10951_CR52","unstructured":"Louizos C, Ullrich K, Welling M (2017) Bayesian compression for deep learning. In: NeurIPS, pp 3288\u20133298"},{"key":"10951_CR53","doi-asserted-by":"crossref","unstructured":"Rastegari M, Ordonez V, Redmon J, Farhadi A (2016) : Xnor-net: ImageNet classification using binary convolutional neural networks. In: European conference on computer vision (ECCV), pp.525\u2013542","DOI":"10.1007\/978-3-319-46493-0_32"},{"key":"10951_CR54","doi-asserted-by":"crossref","unstructured":"Ding X, Ding G, Han J, Tang S (2018) : Auto-balanced filter pruning for efficient convolutional neural networks. In: Association for the Advance of Artificial Intelligence (AAAI), pp.\u00a06665\u20136673","DOI":"10.1609\/aaai.v32i1.12262"},{"key":"10951_CR55","unstructured":"Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H (2017) : MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861,"},{"key":"10951_CR56","doi-asserted-by":"crossref","unstructured":"Ma N, Zhang X, Zheng HT, Sun J (2018) : Shufflenet v2: Practical guidelines for efficient CNN architecture design. In: European conference on computer vision (ECCV), pp.122\u2013138","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"10951_CR57","unstructured":"Tan M, Le Q (2019) : Efficientnet: Rethinking model scaling for convolutional neural networks. In: International Conference on Machine Learning (ICML), pp.6105\u20136114"},{"key":"10951_CR58","unstructured":"Tan M, Le QV (2019) : Mixconv: Mixed depthwise convolutional kernels. In: British Machine Vision Conference (BMVC), pp.\u00a0275\u2013283"},{"issue":"4","key":"10951_CR59","doi-asserted-by":"publisher","first-page":"669","DOI":"10.1142\/S0218001493000339","volume":"7","author":"J Bromley","year":"1993","unstructured":"Bromley J, Guyon I, Lecun Y, Sackinger E, Shah R (1993) Signature verification using a siamese time delay neural network. Pattern Recognit Artif Intell 7(4):669\u2013688","journal-title":"Pattern Recognit Artif Intell"},{"key":"10951_CR60","unstructured":"Kingma D, Ba J (2015) : Adam: A method for stochastic optimization. In International Conference on Learning Representations (ICLR), pp.\u00a0492\u2013507"},{"key":"10951_CR61","doi-asserted-by":"crossref","unstructured":"Xie S, Girshick R, Doll\u00b4 ar P, Tu Z, He K (2017) : Aggregated residual transformations for deep neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a01492\u20131500","DOI":"10.1109\/CVPR.2017.634"},{"key":"10951_CR62","unstructured":"Gooogle (2020) Visual tracker benchmark, Website, http:\/\/cvlab.hanyang.ac.kr\/tracker_benchmark\/datasets.html"},{"key":"10951_CR64","doi-asserted-by":"crossref","unstructured":"Bertinetto L, Valmadre J, Golodetz S et al (2016) : Staple: complementary learners for real-time tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a01401\u20131409","DOI":"10.1109\/CVPR.2016.156"},{"key":"10951_CR65","doi-asserted-by":"crossref","unstructured":"Valmadre J, Bertinetto L, Henriques J et al (2017) : End-to-end representation learning for correlation filter based tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a02805\u20132813","DOI":"10.1109\/CVPR.2017.531"},{"key":"10951_CR66","doi-asserted-by":"crossref","unstructured":"Guo Q, Feng W, Zhou C, Huang R, Wan L, Wang S (2017) : Learning dynamic siamese network for visual object tracking. In: IEEE International Conference on Computer Vision (ICCV), pp.1491 \u2013 1420","DOI":"10.1109\/ICCV.2017.196"},{"key":"10951_CR67","doi-asserted-by":"crossref","unstructured":"Yu Y, Xiong Y, Huang W, Scott MR (2020) : Deformable siamese attention networks for visual object tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a06727\u20136736","DOI":"10.1109\/CVPR42600.2020.00676"}],"updated-by":[{"DOI":"10.1007\/s11063-022-11029-8","type":"correction","label":"Correction","source":"publisher","updated":{"date-parts":[[2022,11,16]],"date-time":"2022-11-16T00:00:00Z","timestamp":1668556800000}}],"container-title":["Neural Processing Letters"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-022-10951-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11063-022-10951-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-022-10951-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,4,26]],"date-time":"2023-04-26T22:34:57Z","timestamp":1682548497000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11063-022-10951-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,8,13]]},"references-count":65,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2023,4]]}},"alternative-id":["10951"],"URL":"https:\/\/doi.org\/10.1007\/s11063-022-10951-1","relation":{},"ISSN":["1370-4621","1573-773X"],"issn-type":[{"value":"1370-4621","type":"print"},{"value":"1573-773X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,8,13]]},"assertion":[{"value":"20 March 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 June 2022","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 June 2022","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 August 2022","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 November 2022","order":5,"name":"change_date","label":"Change Date","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Correction","order":6,"name":"change_type","label":"Change Type","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"A Correction to this paper has been published:","order":7,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"https:\/\/doi.org\/10.1007\/s11063-022-11029-8","URL":"https:\/\/doi.org\/10.1007\/s11063-022-11029-8","order":8,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}