{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,18]],"date-time":"2026-01-18T03:46:55Z","timestamp":1768708015115,"version":"3.49.0"},"reference-count":42,"publisher":"Springer Science and Business Media LLC","issue":"8","license":[{"start":{"date-parts":[[2023,7,22]],"date-time":"2023-07-22T00:00:00Z","timestamp":1689984000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,7,22]],"date-time":"2023-07-22T00:00:00Z","timestamp":1689984000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62066047"],"award-info":[{"award-number":["62066047"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Process Lett"],"published-print":{"date-parts":[[2023,12]]},"DOI":"10.1007\/s11063-023-11365-3","type":"journal-article","created":{"date-parts":[[2023,7,22]],"date-time":"2023-07-22T02:01:39Z","timestamp":1689991299000},"page":"11073-11087","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["RGBT Tracking via Multi-stage Matching Guidance and Context integration"],"prefix":"10.1007","volume":"55","author":[{"given":"Kaixiang","family":"Yan","sequence":"first","affiliation":[]},{"given":"Changcheng","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Dongming","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Ziwei","family":"Zhou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,7,22]]},"reference":[{"key":"11365_CR1","doi-asserted-by":"crossref","unstructured":"Nam H, Han B (2016) Learning multi-domain convolutional neural networks for visual tracking. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4293\u20134302","DOI":"10.1109\/CVPR.2016.465"},{"key":"11365_CR2","unstructured":"Kristan M, Matas J, Leonardis A, Felsberg M, Cehovin L, Fernandez G, Vojir T, Hager G, Nebehay G, Pflugfelder R (2015) The visual object tracking vot2015 challenge results. In: Proceedings of the IEEE international conference on computer vision workshops, pp 1\u201323"},{"key":"11365_CR3","doi-asserted-by":"crossref","unstructured":"Bertinetto L, Valmadre J, Henriques JF, Vedaldi A, Torr PH (2016) Fully-convolutional siamese networks for object tracking. In: European conference on computer vision, pp. 850\u2013865. Springer","DOI":"10.1007\/978-3-319-48881-3_56"},{"key":"11365_CR4","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2019.106977","volume":"96","author":"C Li","year":"2019","unstructured":"Li C, Liang X, Lu Y, Zhao N, Tang J (2019) Rgb-t object tracking: benchmark and baseline. Pattern Recogn 96:106977","journal-title":"Pattern Recogn"},{"issue":"5","key":"11365_CR5","doi-asserted-by":"publisher","first-page":"1562","DOI":"10.1109\/TPAMI.2019.2957464","volume":"43","author":"L Huang","year":"2019","unstructured":"Huang L, Zhao X, Huang K (2019) Got-10k: a large high-diversity benchmark for generic object tracking in the wild. IEEE Trans Pattern Anal Mach Intell 43(5):1562\u20131577","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"11365_CR6","doi-asserted-by":"crossref","unstructured":"Mueller M, Smith N, Ghanem B (2016) A benchmark and simulator for uav tracking. In: European conference on computer vision, pp 445\u2013461. Springer","DOI":"10.1007\/978-3-319-46448-0_27"},{"issue":"15","key":"11365_CR7","doi-asserted-by":"publisher","first-page":"16915","DOI":"10.1109\/JSEN.2021.3078455","volume":"21","author":"J Mei","year":"2021","unstructured":"Mei J, Zhou D, Cao J, Nie R, Guo Y (2021) Hdinet: hierarchical dual-sensor interaction network for RGBT tracking. IEEE Sens J 21(15):16915\u201316926","journal-title":"IEEE Sens J"},{"key":"11365_CR8","doi-asserted-by":"publisher","first-page":"567","DOI":"10.1109\/tmm.2021.3055362","volume":"24","author":"Q Xu","year":"2022","unstructured":"Xu Q, Mei YM, Liu JP, Li CL (2022) Multimodal cross-layer bilinear pooling for RGBT tracking. IEEE Trans Multimed 24:567\u2013580. https:\/\/doi.org\/10.1109\/tmm.2021.3055362","journal-title":"IEEE Trans Multimed"},{"issue":"7","key":"11365_CR9","doi-asserted-by":"publisher","first-page":"7084","DOI":"10.1109\/jsen.2022.3154657","volume":"22","author":"YD Li","year":"2022","unstructured":"Li YD, Lai HC, Wang LJ, Jia ZH (2022) Multibranch adaptive fusion network for RGBT tracking. IEEE Sens J 22(7):7084\u20137093. https:\/\/doi.org\/10.1109\/jsen.2022.3154657","journal-title":"IEEE Sens J"},{"key":"11365_CR10","doi-asserted-by":"publisher","DOI":"10.1109\/TIM.2022.3193971","author":"J Mei","year":"2022","unstructured":"Mei J, Liu Y, Wang C, Zhou D, Nie R (2022) Cao J (2022) Asymmetric global-local mutual integration network for RGBT tracking. IEEE Trans Instrum Meas. https:\/\/doi.org\/10.1109\/TIM.2022.3193971","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"11365_CR11","doi-asserted-by":"publisher","first-page":"327","DOI":"10.1016\/j.neucom.2022.04.017","volume":"493","author":"W Xia","year":"2022","unstructured":"Xia W, Zhou D, Cao J, Liu Y, Hou R (2022) Cirnet: an improved RGBT tracking via cross-modality interaction and re-identification. Neurocomputing 493:327\u2013339. https:\/\/doi.org\/10.1016\/j.neucom.2022.04.017","journal-title":"Neurocomputing"},{"key":"11365_CR12","unstructured":"Zhu Y, Li C, Luo B, Tang J, Wang X Dense feature aggregation and pruning for rgbt tracking. In: Proceedings of the 27th ACM international conference on multimedia, pp 465\u2013472"},{"key":"11365_CR13","unstructured":"Gao Y, Li C, Zhu Y, Tang J, He T, Wang F Deep adaptive fusion network for high performance RGBT tracking. In: Proceedings of the IEEE\/CVF international conference on computer vision workshops, pp 0\u20130"},{"issue":"3","key":"11365_CR14","doi-asserted-by":"publisher","first-page":"3276","DOI":"10.1007\/s10489-021-02542-9","volume":"52","author":"ZB Su","year":"2022","unstructured":"Su ZB, Li W, Ma Z, Gao R (2022) An improved u-net method for the semantic segmentation of remote sensing images. Appl Intell 52(3):3276\u20133288. https:\/\/doi.org\/10.1007\/s10489-021-02542-9","journal-title":"Appl Intell"},{"key":"11365_CR15","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1016\/j.patrec.2021.07.013","volume":"150","author":"XY Liu","year":"2021","unstructured":"Liu XY, Sang J, Wu WQ, Liu K, Liu Q, Xia XF (2021) Density-aware and background-aware network for crowd counting via multi-task learning. Pattern Recogn Lett 150:221\u2013227. https:\/\/doi.org\/10.1016\/j.patrec.2021.07.013","journal-title":"Pattern Recogn Lett"},{"issue":"5","key":"11365_CR16","doi-asserted-by":"publisher","first-page":"2019","DOI":"10.1109\/tip.2014.2311377","volume":"23","author":"J Yu","year":"2014","unstructured":"Yu J, Rui Y, Tao DC (2014) Click prediction for web image reranking using multimodal sparse coding. IEEE Trans Image Process 23(5):2019\u20132032. https:\/\/doi.org\/10.1109\/tip.2014.2311377","journal-title":"IEEE Trans Image Process"},{"issue":"2","key":"11365_CR17","doi-asserted-by":"publisher","first-page":"563","DOI":"10.1109\/tpami.2019.2932058","volume":"44","author":"J Yu","year":"2022","unstructured":"Yu J, Tan M, Zhang HY, Tao DC, Rui Y (2022) Hierarchical deep click feature prediction for fine-grained image recognition. IEEE Trans Pattern Anal Mach Intell 44(2):563\u2013578. https:\/\/doi.org\/10.1109\/tpami.2019.2932058","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"11365_CR18","doi-asserted-by":"publisher","first-page":"7790","DOI":"10.1109\/tip.2021.3109518","volume":"30","author":"WJ Zhou","year":"2021","unstructured":"Zhou WJ, Liu JF, Lei JS, Yu L, Hwang JN (2021) Gmnet: graded-feature multilabel-learning network for RGB-thermal urban scene semantic segmentation. IEEE Trans Image Process 30:7790\u20137802. https:\/\/doi.org\/10.1109\/tip.2021.3109518","journal-title":"IEEE Trans Image Process"},{"issue":"12","key":"11365_CR19","doi-asserted-by":"publisher","first-page":"5743","DOI":"10.1109\/TIP.2016.2614135","volume":"25","author":"C Li","year":"2016","unstructured":"Li C, Cheng H, Hu S, Liu X, Tang J, Lin L (2016) Learning collaborative sparse representation for grayscale-thermal tracking. IEEE Trans Image Process 25(12):5743\u20135756","journal-title":"IEEE Trans Image Process"},{"key":"11365_CR20","doi-asserted-by":"crossref","unstructured":"Bhat G, Danelljan M, Gool LV, Timofte R (2019) Learning discriminative model prediction for tracking. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 6182\u20136191","DOI":"10.1109\/ICCV.2019.00628"},{"key":"11365_CR21","doi-asserted-by":"publisher","first-page":"189","DOI":"10.1016\/j.knosys.2017.07.032","volume":"134","author":"Q Liu","year":"2017","unstructured":"Liu Q, Lu XH, He ZY, Zhang CK, Chen WS (2017) Deep convolutional neural networks for thermal infrared object tracking. Knowl-Based Syst 134:189\u2013198. https:\/\/doi.org\/10.1016\/j.knosys.2017.07.032","journal-title":"Knowl-Based Syst"},{"key":"11365_CR22","doi-asserted-by":"publisher","unstructured":"Li B, Yan JJ, Wu W, Zhu Z, Hu XL (2018) Ieee: High performance visual tracking with siamese region proposal network. In: 31st IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Conference on Computer Vision and Pattern Recognition, pp. 8971\u20138980. Ieee, New York. https:\/\/doi.org\/10.1109\/cvpr.2018.00935.<Go to ISI>:\/\/WOS:000457843609015","DOI":"10.1109\/cvpr.2018.00935"},{"key":"11365_CR23","doi-asserted-by":"publisher","unstructured":"Valmadre J, Bertinetto L, Henriques J, Vedaldi A, Torr PHS (2017) Ieee: End-to-end representation learning for correlation filter based tracking. In: 30th IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE conference on computer vision and pattern recognition, pp. 5000\u20135008. Ieee, New York. https:\/\/doi.org\/10.1109\/cvpr.2017.531.<Go to ISI>:\/\/WOS:000418371405010","DOI":"10.1109\/cvpr.2017.531"},{"key":"11365_CR24","doi-asserted-by":"publisher","unstructured":"Chen ZD, Zhong BN, Li GR, Zhang SP, Ji RR (2020) Ieee: siamese box adaptive network for visual tracking. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE conference on computer vision and pattern recognition, pp 6667\u20136676. IEEE, New York. https:\/\/doi.org\/10.1109\/cvpr42600.2020.00670.<Go to ISI>:\/\/WOS:000620679506094","DOI":"10.1109\/cvpr42600.2020.00670"},{"issue":"8","key":"11365_CR25","doi-asserted-by":"publisher","first-page":"1561","DOI":"10.1109\/TPAMI.2016.2609928","volume":"39","author":"M Danelljan","year":"2017","unstructured":"Danelljan M, Hager G, Khan FS, Felsberg M (2017) Discriminative scale space tracking. IEEE Trans Pattern Anal Mach Intell 39(8):1561\u20131575. https:\/\/doi.org\/10.1109\/TPAMI.2016.2609928","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"5","key":"11365_CR26","doi-asserted-by":"publisher","first-page":"1500","DOI":"10.1049\/ipr2.12427","volume":"16","author":"XH Liu","year":"2022","unstructured":"Liu XH, Luo YC, Yan KD, Chen JF, Lei ZY (2022) Cmc2r: Cross-modal collaborative contextual representation for RGBT tracking. IET Image Proc 16(5):1500\u20131510. https:\/\/doi.org\/10.1049\/ipr2.12427","journal-title":"IET Image Proc"},{"key":"11365_CR27","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2022.3157594","author":"A Lu","year":"2022","unstructured":"Lu A, Qian C, Li C, Tang J, Wang L (2022) Duality-gated mutual condition network for RGBT tracking. IEEE Trans Neural Netw Learn Syst. https:\/\/doi.org\/10.1109\/TNNLS.2022.3157594","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"issue":"1","key":"11365_CR28","doi-asserted-by":"publisher","first-page":"121","DOI":"10.1109\/TIV.2020.2980735","volume":"6","author":"Y Zhu","year":"2020","unstructured":"Zhu Y, Li C, Tang J, Luo B (2020) Quality-aware feature aggregation network for robust RGBT tracking. IEEE Trans Intell Veh 6(1):121\u2013130","journal-title":"IEEE Trans Intell Veh"},{"key":"11365_CR29","unstructured":"Long\u00a0Li C, Lu A, Hua\u00a0Zheng A, Tu Z, Tang J (2019) Multi-adapter rgbt tracking. In: Proceedings of the IEEE\/CVF international conference on computer vision workshops, pp 0\u20130"},{"key":"11365_CR30","doi-asserted-by":"crossref","unstructured":"Wang C, Xu C, Cui Z, Zhou L, Zhang T, Zhang X, Yang J (2020) Cross-modal pattern-propagation for rgb-t tracking. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 7064\u20137073","DOI":"10.1109\/CVPR42600.2020.00709"},{"issue":"2","key":"11365_CR31","doi-asserted-by":"publisher","first-page":"579","DOI":"10.1109\/TCSVT.2021.3067997","volume":"32","author":"Y Zhu","year":"2021","unstructured":"Zhu Y, Li C, Tang J, Luo B, Wang L (2021) Rgbt tracking by trident fusion network. IEEE Trans Circuits Syst Video Technol 32(2):579\u2013592","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"11365_CR32","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.patcog.2021.107952","volume":"116","author":"J Zhang","year":"2021","unstructured":"Zhang J, Cao YY, Wu Q (2021) Vector of locally and adaptively aggregated descriptors for image feature representation. Pattern Recogn 116:11. https:\/\/doi.org\/10.1016\/j.patcog.2021.107952","journal-title":"Pattern Recogn"},{"issue":"5","key":"11365_CR33","doi-asserted-by":"publisher","first-page":"3117","DOI":"10.1002\/int.22814","volume":"37","author":"J Zhang","year":"2022","unstructured":"Zhang J, Yang JN, Yu J, Fan JP (2022) Semisupervised image classification by mutual learning of multiple self-supervised models. Int J Intell Syst 37(5):3117\u20133141. https:\/\/doi.org\/10.1002\/int.22814","journal-title":"Int J Intell Syst"},{"key":"11365_CR34","unstructured":"Xu D, Ouyang WL, Alameda-Pineda X, Ricci E, Wang XG, Sebe N (2017) Learning deep structured multi-scale features using attention-gated crfs for contour prediction. In: 31st Annual Conference on Neural Information Processing Systems (NIPS). Advances in Neural Information Processing Systems, vol. 30. Neural Information Processing Systems (Nips), LA JOLLA.<Go to ISI>:\/\/WOS:000452649404004"},{"key":"11365_CR35","doi-asserted-by":"publisher","unstructured":"Fu JL, Zheng HL, Mei T (2017) Ieee: Look closer to see better: Recurrent attention convolutional neural network for fine-grained image recognition. In: 30th IEEE\/CVF conference on computer vision and pattern recognition (CVPR). pp 4476\u20134484. Ieee, New York. https:\/\/doi.org\/10.1109\/cvpr.2017.476.<Go to ISI>:\/\/WOS:000418371404060","DOI":"10.1109\/cvpr.2017.476"},{"key":"11365_CR36","unstructured":"Long\u00a0Li C, Lu A, Hua\u00a0Zheng A, Tu Z, Tang J (2019) Multi-adapter RGBT tracking. In: Proceedings of the IEEE\/CVF international conference on computer vision workshops, pp 0"},{"key":"11365_CR37","unstructured":"Chen P, Gao J, Yuan Y, Wang Q (2022) Mafnet: a multi-attention fusion network for rgb-t crowd counting. arXiv preprint arXiv:2208.06761"},{"key":"11365_CR38","doi-asserted-by":"crossref","unstructured":"Wang Q, Teng Z, Xing J, Gao J, Hu W, Maybank S (2018) Learning attentions: residual attentional siamese network for high performance online visual tracking. In: Proceedings of the ieee conference on computer vision and pattern recognition, pp 4854\u20134863","DOI":"10.1109\/CVPR.2018.00510"},{"key":"11365_CR39","doi-asserted-by":"publisher","first-page":"5613","DOI":"10.1109\/TIP.2021.3087341","volume":"30","author":"A Lu","year":"2021","unstructured":"Lu A, Li C, Yan Y, Tang J, Luo B (2021) Rgbt tracking via multi-adapter network with hierarchical divergence loss. IEEE Trans Image Process 30:5613\u20135625","journal-title":"IEEE Trans Image Process"},{"key":"11365_CR40","doi-asserted-by":"publisher","first-page":"3335","DOI":"10.1109\/TIP.2021.3060862","volume":"30","author":"P Zhang","year":"2021","unstructured":"Zhang P, Zhao J, Bo C, Wang D, Lu H, Yang X (2021) Jointly modeling motion and appearance cues for robust rgb-t tracking. IEEE Trans Image Process 30:3335\u20133347","journal-title":"IEEE Trans Image Process"},{"key":"11365_CR41","unstructured":"Danelljan M, Bhat G, Shahbaz\u00a0Khan F, Felsberg M Eco: Efficient convolution operators for tracking. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6638\u20136646"},{"key":"11365_CR42","doi-asserted-by":"crossref","unstructured":"Li C, Zhao N, Lu Y, Zhu C, Tang J (2017) Weighted sparse representation regularized graph learning for rgb-t object tracking. In: Proceedings of the 25th ACM international conference on multimedia, pp 1856\u20131864","DOI":"10.1145\/3123266.3123289"}],"container-title":["Neural Processing Letters"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-023-11365-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11063-023-11365-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-023-11365-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,22]],"date-time":"2023-11-22T05:08:51Z","timestamp":1700629731000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11063-023-11365-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,7,22]]},"references-count":42,"journal-issue":{"issue":"8","published-print":{"date-parts":[[2023,12]]}},"alternative-id":["11365"],"URL":"https:\/\/doi.org\/10.1007\/s11063-023-11365-3","relation":{},"ISSN":["1370-4621","1573-773X"],"issn-type":[{"value":"1370-4621","type":"print"},{"value":"1573-773X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,7,22]]},"assertion":[{"value":"10 July 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 July 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"To the best of our knowledge, the named authors have no conflict of interest, financial or otherwise. The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing Interest"}}]}}