{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,1]],"date-time":"2025-10-01T15:24:30Z","timestamp":1759332270309,"version":"3.37.3"},"reference-count":75,"publisher":"Springer Science and Business Media LLC","issue":"9-11","license":[{"start":{"date-parts":[[2021,8,9]],"date-time":"2021-08-09T00:00:00Z","timestamp":1628467200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,8,9]],"date-time":"2021-08-09T00:00:00Z","timestamp":1628467200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61972067","91748104"],"award-info":[{"award-number":["61972067","91748104"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Innovation Technology Funding of Dalian","award":["2018J11CY010","2020JJ26GX036"],"award-info":[{"award-number":["2018J11CY010","2020JJ26GX036"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2021,9]]},"DOI":"10.1007\/s00371-021-02237-9","type":"journal-article","created":{"date-parts":[[2021,8,9]],"date-time":"2021-08-09T17:26:02Z","timestamp":1628529962000},"page":"2671-2683","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["Multi-domain collaborative feature representation for robust visual object tracking"],"prefix":"10.1007","volume":"37","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0061-5465","authenticated-orcid":false,"given":"Jiqing","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Kai","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Bo","family":"Dong","sequence":"additional","affiliation":[]},{"given":"Yingkai","family":"Fu","sequence":"additional","affiliation":[]},{"given":"Yuxin","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Baocai","family":"Yin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,8,9]]},"reference":[{"key":"2237_CR1","doi-asserted-by":"crossref","unstructured":"Barranco, F., Fermuller, C., Ros, E.: Real-time clustering and multi-target tracking using event-based sensors. In: IEEE\/RSJ International Conference on Intelligent Robots and Systems (2018)","DOI":"10.1109\/IROS.2018.8593380"},{"key":"2237_CR2","doi-asserted-by":"crossref","unstructured":"Bertinetto, L., Valmadre, J., Henriques, J.F., Vedaldi, A., Torr, P.H.: Fully-convolutional siamese networks for object tracking. In: Proceedings of the European Conference on Computer Vision. Springer (2016)","DOI":"10.1007\/978-3-319-48881-3_56"},{"key":"2237_CR3","doi-asserted-by":"crossref","unstructured":"Bhat, G., Danelljan, M., Gool, L.V., Timofte, R.: Learning discriminative model prediction for tracking. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (2019)","DOI":"10.1109\/ICCV.2019.00628"},{"key":"2237_CR4","doi-asserted-by":"crossref","unstructured":"Bi, Y., Chadha, A., Abbas, A., Bourtsoulatze, E., Andreopoulos, Y.: Graph-based object classification for neuromorphic vision sensing. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (2019)","DOI":"10.1109\/ICCV.2019.00058"},{"key":"2237_CR5","doi-asserted-by":"crossref","unstructured":"Br\u00e4ndli, C., Berner, R., Yang, M., Liu, S.C., Delbr\u00fcck, T.: A 240 $$\\times $$ 180 130 db 3 $$\\mu $$s latency global shutter spatiotemporal vision sensor. IEEE J. Solid-state Circ. (2014)","DOI":"10.1109\/JSSC.2014.2342715"},{"key":"2237_CR6","doi-asserted-by":"crossref","unstructured":"Cadena, P.R.G., Qian, Y., Wang, C., Yang, M.: Spade-e2vid: Spatially-adaptive denormalization for event-based video reconstruction. IEEE Trans. Image Process. (2021)","DOI":"10.1109\/TIP.2021.3052070"},{"key":"2237_CR7","doi-asserted-by":"crossref","unstructured":"Chen, H., Suter, D., Wu, Q., Wang, H.: End-to-end learning of object motion estimation from retinal events for event-based object tracking. In: Proceedings of the AAAI Conference on Artificial Intelligence (2020)","DOI":"10.1609\/aaai.v34i07.6625"},{"key":"2237_CR8","doi-asserted-by":"crossref","unstructured":"Chen, H., Wu, Q., Liang, Y., Gao, X., Wang, H.: Asynchronous tracking-by-detection on adaptive time surfaces for event-based object tracking. In: Proceedings of the 27th ACM International Conference on Multimedia (2019)","DOI":"10.1145\/3343031.3350975"},{"key":"2237_CR9","unstructured":"Choi, J., Yoon, K.J., et\u00a0al.: Learning to super resolve intensity images from events. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2020)"},{"key":"2237_CR10","doi-asserted-by":"crossref","unstructured":"Cohen, G.K., Orchard, G., Leng, S.H., Tapson, J., Benosman, R.B., Van\u00a0Schaik, A.: Skimming digits: neuromorphic classification of spike-encoded images. Front. Neurosci. (2016)","DOI":"10.3389\/fnins.2016.00184"},{"key":"2237_CR11","doi-asserted-by":"crossref","unstructured":"Dai, K., Wang, D., Lu, H., Sun, C., Li, J.: Visual tracking via adaptive spatially-regularized correlation filters. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2019)","DOI":"10.1109\/CVPR.2019.00480"},{"key":"2237_CR12","doi-asserted-by":"crossref","unstructured":"Danelljan, M., Bhat, G., Khan, F.S., Felsberg, M.: Atom: Accurate tracking by overlap maximization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2019)","DOI":"10.1109\/CVPR.2019.00479"},{"key":"2237_CR13","doi-asserted-by":"crossref","unstructured":"Danelljan, M., Bhat, G., Shahbaz\u00a0Khan, F., Felsberg, M.: Eco: Efficient convolution operators for tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2017)","DOI":"10.1109\/CVPR.2017.733"},{"key":"2237_CR14","doi-asserted-by":"crossref","unstructured":"Danelljan, M., Hager, G., Shahbaz\u00a0Khan, F., Felsberg, M.: Learning spatially regularized correlation filters for visual tracking. In: Proceedings of the IEEE International Conference on Computer Vision (2015)","DOI":"10.1109\/ICCV.2015.490"},{"key":"2237_CR15","doi-asserted-by":"crossref","unstructured":"Fan, H., Ling, H.: Siamese cascaded region proposal networks for real-time visual tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2019)","DOI":"10.1109\/CVPR.2019.00814"},{"key":"2237_CR16","doi-asserted-by":"crossref","unstructured":"Gehrig, D., Loquercio, A., Derpanis, K.G., Scaramuzza, D.: End-to-end learning of representations for asynchronous event-based data. In: Proceedings of the IEEE International Conference on Computer Vision (2019)","DOI":"10.1109\/ICCV.2019.00573"},{"key":"2237_CR17","doi-asserted-by":"crossref","unstructured":"Gehrig, M., Shrestha, S.B., Mouritzen, D., Scaramuzza, D.: Event-based angular velocity regression with spiking networks. In: 2020 IEEE International Conference on Robotics and Automation (ICRA) (2020)","DOI":"10.1109\/ICRA40945.2020.9197133"},{"key":"2237_CR18","doi-asserted-by":"publisher","first-page":"738","DOI":"10.1103\/PhysRevE.51.738","volume":"51","author":"W Gerstner","year":"1995","unstructured":"Gerstner, W.: Time structure of the activity in neural network models. Phys. Rev. E 51, 738\u2013758 (1995)","journal-title":"Phys. Rev. E"},{"key":"2237_CR19","doi-asserted-by":"crossref","unstructured":"He, A., Luo, C., Tian, X., Zeng, W.: A twofold siamese network for real-time object tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2018)","DOI":"10.1109\/CVPR.2018.00508"},{"key":"2237_CR20","doi-asserted-by":"publisher","first-page":"583","DOI":"10.1109\/TPAMI.2014.2345390","volume":"37","author":"JF Henriques","year":"2014","unstructured":"Henriques, J.F., Caseiro, R., Martins, P., Batista, J.: High-speed tracking with kernelized correlation filters. IEEE Trans. Pattern Anal. Mach. Intell. 37, 583\u2013596 (2014)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2237_CR21","doi-asserted-by":"publisher","first-page":"1562","DOI":"10.1109\/TPAMI.2019.2957464","volume":"43","author":"L Huang","year":"2019","unstructured":"Huang, L., Zhao, X., Huang, K.: Got-10k: A large high-diversity benchmark for generic object tracking in the wild. IEEE Trans. Pattern Anal. Mach. Intell. 43, 1562\u20131577 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2237_CR22","doi-asserted-by":"crossref","unstructured":"Jung, I., Son, J., Baek, M., Han, B.: Real-time mdnet. In: Proceedings of the European Conference on Computer Vision (2018)","DOI":"10.1007\/978-3-030-01225-0_6"},{"key":"2237_CR23","doi-asserted-by":"crossref","unstructured":"Kalal, Z., Mikolajczyk, K., Matas, J.: Tracking-learning-detection. IEEE Trans. Pattern Anal. Mach. Intell. (2011)","DOI":"10.1109\/TPAMI.2011.239"},{"key":"2237_CR24","doi-asserted-by":"crossref","unstructured":"Kart, U., K\u00e4m\u00e4r\u00e4inen, J.K., Matas, J., Fan, L., Cricri, F.: Depth masked discriminative correlation filter. In: 2018 24th International Conference on Pattern Recognition (2018)","DOI":"10.1109\/ICPR.2018.8546179"},{"key":"2237_CR25","doi-asserted-by":"crossref","unstructured":"Kart, U., K\u00e4m\u00e4r\u00e4inen, J.K., Matas, J., Matas, J.: How to make an rgbd tracker? In: Proceedings of the European Conference on Computer Vision (2018)","DOI":"10.1007\/978-3-030-11009-3_8"},{"key":"2237_CR26","doi-asserted-by":"crossref","unstructured":"Kepple, D.R., Lee, D., Prepsius, C., Isler, V., Park, I.M., Lee, D.D.: Jointly learning visual motion and confidence from local patches in event cameras. In: Proceedings of the European Conference on Computer Vision (2020)","DOI":"10.1007\/978-3-030-58539-6_30"},{"key":"2237_CR27","doi-asserted-by":"crossref","unstructured":"Lan, X., Ye, M., Zhang, S., Yuen, P.C.: Robust collaborative discriminative learning for rgb-infrared tracking. In: Thirty-Second AAAI Conference on Artificial Intelligence (2018)","DOI":"10.1609\/aaai.v32i1.12307"},{"key":"2237_CR28","doi-asserted-by":"crossref","unstructured":"Li, B., Wu, W., Wang, Q., Zhang, F., Xing, J., Yan, J.: Siamrpn++: Evolution of siamese visual tracking with very deep networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2019)","DOI":"10.1109\/CVPR.2019.00441"},{"key":"2237_CR29","doi-asserted-by":"crossref","unstructured":"Li, B., Yan, J., Wu, W., Zhu, Z., Hu, X.: High performance visual tracking with siamese region proposal network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2018)","DOI":"10.1109\/CVPR.2018.00935"},{"key":"2237_CR30","doi-asserted-by":"crossref","unstructured":"Li, C., Lu, A., Zheng, A., Tu, Z., Tang, J.: Multi-adapter rgbt tracking. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision Workshops (2019)","DOI":"10.1109\/ICCVW.2019.00279"},{"key":"2237_CR31","doi-asserted-by":"crossref","unstructured":"Li, C., Zhu, C., Huang, Y., Tang, J., Wang, L.: Cross-modal ranking with soft consistency and noisy labels for robust rgb-t tracking. In: Proceedings of the European Conference on Computer Vision (2018)","DOI":"10.1007\/978-3-030-01261-8_49"},{"key":"2237_CR32","doi-asserted-by":"crossref","unstructured":"Li, P., Chen, B., Ouyang, W., Wang, D., Yang, X., Lu, H.: Gradnet: Gradient-guided network for visual object tracking. In: Proceedings of the IEEE International Conference on Computer Vision (2019)","DOI":"10.1109\/ICCV.2019.00626"},{"key":"2237_CR33","doi-asserted-by":"crossref","unstructured":"Li, W., Li, X., Bourahla, O.E., Huang, F., Wu, F., Liu, W., Wang, Z., Liu, H.: Progressive multistage learning for discriminative tracking. IEEE Trans. Cybern. (2020)","DOI":"10.1109\/TCYB.2020.2985398"},{"key":"2237_CR34","doi-asserted-by":"crossref","unstructured":"Mei, H., Liu, Y., Wei, Z., Zhou, D., Xiaopeng, X., Zhang, Q., Yang, X.: Exploring dense context for salient object detection. IEEE Trans. Circuits Syst. Video Technol. (2021)","DOI":"10.1109\/TCSVT.2021.3069848"},{"key":"2237_CR35","doi-asserted-by":"crossref","unstructured":"Mei, H., Yang, X., Wang, Y., Liu, Y., He, S., Zhang, Q., Wei, X., Lau, R.W.: Don\u2019t hit me! glass detection in real-world scenes. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2020)","DOI":"10.1109\/CVPR42600.2020.00374"},{"key":"2237_CR36","doi-asserted-by":"crossref","unstructured":"Mitrokhin, A., Fermuller, C., Parameshwara, C., Aloimonos, Y.: Event-based moving object detection and tracking. In: IEEE\/RSJ International Conference on Intelligent Robots and Systems (2018)","DOI":"10.1109\/IROS.2018.8593805"},{"key":"2237_CR37","doi-asserted-by":"crossref","unstructured":"Mostafavi, M., Wang, L., Yoon, K.J.: Learning to reconstruct hdr images from events, with applications to depth and flow prediction. Int. J. Comput. Vis. (2021)","DOI":"10.1007\/s11263-020-01410-2"},{"key":"2237_CR38","doi-asserted-by":"crossref","unstructured":"Nam, H., Han, B.: Learning multi-domain convolutional neural networks for visual tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016)","DOI":"10.1109\/CVPR.2016.465"},{"key":"2237_CR39","doi-asserted-by":"publisher","first-page":"51","DOI":"10.1109\/MSP.2019.2931595","volume":"36","author":"EO Neftci","year":"2019","unstructured":"Neftci, E.O., Mostafa, H., Zenke, F.: Surrogate gradient learning in spiking neural networks: Bringing the power of gradient-based optimization to spiking neural networks. IEEE Signal Process. Mag. 36, 51\u201363 (2019)","journal-title":"IEEE Signal Process. Mag."},{"key":"2237_CR40","doi-asserted-by":"crossref","unstructured":"Pan, L., Liu, M., Hartley, R.: Single image optical flow estimation with an event camera. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2020)","DOI":"10.1109\/CVPR42600.2020.00174"},{"key":"2237_CR41","doi-asserted-by":"crossref","unstructured":"Piatkowska, E., Belbachir, A.N., Schraml, S., Gelautz, M.: Spatiotemporal multiple persons tracking using dynamic vision sensor. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (2012)","DOI":"10.1109\/CVPRW.2012.6238892"},{"key":"2237_CR42","doi-asserted-by":"crossref","unstructured":"Qiao, Y., Liu, Y., Yang, X., Zhou, D., Xu, M., Zhang, Q., Wei, X.: Attention-guided hierarchical structure aggregation for image matting. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2020)","DOI":"10.1109\/CVPR42600.2020.01369"},{"key":"2237_CR43","doi-asserted-by":"crossref","unstructured":"Qiao, Y., Liu, Y., Zhu, Q., Yang, X., Wang, Y., Zhang, Q., Wei, X.: Multi-scale information assembly for image matting. In: Computer Graphics Forum (2020)","DOI":"10.1111\/cgf.14168"},{"key":"2237_CR44","doi-asserted-by":"crossref","unstructured":"Ramesh, B., Zhang, S., Yang, H., Ussa, A., Ong, M., Orchard, G., Xiang, C.: e-tld: Event-based framework for dynamic object tracking. IEEE Trans. Circuits Syst. Video Technol. (2020)","DOI":"10.1109\/TCSVT.2020.3044287"},{"key":"2237_CR45","unstructured":"Rebecq, H., Gehrig, D., Scaramuzza, D.: Esim: an open event camera simulator. In: Conference on Robot Learning (2018)"},{"key":"2237_CR46","doi-asserted-by":"crossref","unstructured":"Rebecq, H., Ranftl, R., Koltun, V., Scaramuzza, D.: Events-to-video: Bringing modern computer vision to event cameras. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2019)","DOI":"10.1109\/CVPR.2019.00398"},{"key":"2237_CR47","doi-asserted-by":"crossref","unstructured":"Ren, W., Wang, X., Tian, J., Tang, Y., Chan, A.B.: Tracking-by-counting: Using network flows on crowd density maps for tracking multiple targets. IEEE Trans. Image Process. (2020)","DOI":"10.1109\/TIP.2020.3044219"},{"key":"2237_CR48","unstructured":"Shrestha, S.B., Orchard, G.: Slayer: Spike layer error reassignment in time. In: Advances in Neural Information Processing Systems (2018)"},{"key":"2237_CR49","volume-title":"Advances in Neural Information Processing Systems","author":"SB Shrestha","year":"2018","unstructured":"Shrestha, S.B., Orchard, G.: SLAYER: Spike layer error reassignment in time. In: Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R. (eds.) Advances in Neural Information Processing Systems. Curran Associates Inc, Red Hook (2018)"},{"key":"2237_CR50","first-page":"3126","volume":"29","author":"SB Shrestha","year":"2017","unstructured":"Shrestha, S.B., Song, Q.: Robustness to training disturbances in Spikeprop learning. IEEE Trans. Neural Netw. Learn. Syst. 29, 3126\u20133139 (2017)","journal-title":"Learn. Syst."},{"key":"2237_CR51","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"2237_CR52","doi-asserted-by":"crossref","unstructured":"Song, S., Xiao, J.: Tracking revisited using rgbd camera: Unified benchmark and baselines. In: Proceedings of the IEEE International Conference on Computer Vision (2013)","DOI":"10.1109\/ICCV.2013.36"},{"key":"2237_CR53","doi-asserted-by":"crossref","unstructured":"Stoffregen, T., Gallego, G., Drummond, T., Kleeman, L., Scaramuzza, D.: Event-based motion segmentation by motion compensation. In: Proceedings of the IEEE International Conference on Computer Vision (2019)","DOI":"10.1109\/ICCV.2019.00734"},{"key":"2237_CR54","doi-asserted-by":"publisher","first-page":"47","DOI":"10.1016\/j.neunet.2018.12.002","volume":"111","author":"A Tavanaei","year":"2019","unstructured":"Tavanaei, A., Ghodrati, M., Kheradpisheh, S.R., Masquelier, T., Maida, A.: Deep learning in spiking neural networks. Neural Netw 111, 47\u201363 (2019)","journal-title":"Neural Netw"},{"key":"2237_CR55","doi-asserted-by":"crossref","unstructured":"Tulyakov, S., Fleuret, F., Kiefel, M., Gehler, P., Hirsch, M.: Learning an event sequence embedding for dense event-based deep stereo. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (2019)","DOI":"10.1109\/ICCV.2019.00161"},{"key":"2237_CR56","doi-asserted-by":"crossref","unstructured":"Valmadre, J., Bertinetto, L., Henriques, J., Vedaldi, A., Torr, P.H.: End-to-end representation learning for correlation filter based tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2017)","DOI":"10.1109\/CVPR.2017.531"},{"key":"2237_CR57","doi-asserted-by":"crossref","unstructured":"Wang, L., Ho, Y.S., Yoon, K.J., et\u00a0al.: Event-based high dynamic range image and very high frame rate video generation using conditional generative adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2019)","DOI":"10.1109\/CVPR.2019.01032"},{"key":"2237_CR58","doi-asserted-by":"publisher","first-page":"4765","DOI":"10.1109\/TIP.2017.2723239","volume":"26","author":"X Wang","year":"2017","unstructured":"Wang, X., Fan, B., Chang, S., Wang, Z., Liu, X., Tao, D., Huang, T.S.: Greedy batch-based minimum-cost flows for tracking multiple objects. IEEE Trans. Image Process. 26, 4765\u20134776 (2017)","journal-title":"IEEE Trans. Image Process."},{"key":"2237_CR59","doi-asserted-by":"publisher","first-page":"1834","DOI":"10.1109\/TPAMI.2014.2388226","volume":"37","author":"Y Wu","year":"2015","unstructured":"Wu, Y., Lim, J., Yang, M.H.: Object tracking benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 37, 1834\u20131848 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2237_CR60","doi-asserted-by":"crossref","first-page":"2485","DOI":"10.1109\/TCYB.2017.2740952","volume":"48","author":"J Xiao","year":"2017","unstructured":"Xiao, J., Stolkin, R., Gao, Y., Leonardis, A.: Robust fusion of color and depth data for RGB-D target tracking using adaptive range-invariant depth models and spatio-temporal consistency constraints. IEEE Trans. Cybern. 48, 2485\u20132499 (2017)","journal-title":"IEEE Trans. Cybern."},{"key":"2237_CR61","doi-asserted-by":"crossref","unstructured":"Xu, K., Yang, X., Yin, B., Lau, R.W.: Learning to restore low-light images via decomposition-and-enhancement. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2020)","DOI":"10.1109\/CVPR42600.2020.00235"},{"key":"2237_CR62","doi-asserted-by":"crossref","unstructured":"Yang, X., Mei, H., Xu, K., Wei, X., Yin, B., Lau, R.W.: Where is my mirror? In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (2019)","DOI":"10.1109\/ICCV.2019.00890"},{"key":"2237_CR63","doi-asserted-by":"publisher","first-page":"328","DOI":"10.1109\/TMM.2018.2863602","volume":"21","author":"X Yang","year":"2018","unstructured":"Yang, X., Mei, H., Zhang, J., Xu, K., Yin, B., Zhang, Q., Wei, X.: DRFN: Deep recurrent fusion network for single-image super-resolution with large factors. IEEE Trans. Multimedia 21, 328\u2013337D (2018)","journal-title":"IEEE Trans. Multimedia"},{"key":"2237_CR64","unstructured":"Yang, X., Xu, K., Chen, S., He, S., Yin, B.Y., Lau, R.: Active matting. Adv. Neural Inf. Process. Syst. (2018)"},{"key":"2237_CR65","doi-asserted-by":"crossref","unstructured":"Yang, X., Xu, K., Song, Y., Zhang, Q., Wei, X., Lau, R.W.: Image correction via deep reciprocating hdr transformation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2018)","DOI":"10.1109\/CVPR.2018.00193"},{"key":"2237_CR66","doi-asserted-by":"crossref","unstructured":"Zenke, F., Ganguli, S.: Superspike: Supervised learning in multilayer spiking neural networks. Neural Comput. (2018)","DOI":"10.1162\/neco_a_01086"},{"key":"2237_CR67","doi-asserted-by":"crossref","unstructured":"Zhang, J., Long, C., Wang, Y., Piao, H., Mei, H., Yang, X., Yin, B.: A two-stage attentive network for single image super-resolution. IEEE Trans. Circuits Syst. Video Technol. (2021)","DOI":"10.1109\/TCSVT.2021.3071191"},{"key":"2237_CR68","doi-asserted-by":"crossref","unstructured":"Zhang, J., Long, C., Wang, Y., Yang, X., Mei, H., Yin, B.: Multi-context and enhanced reconstruction network for single image super resolution. In: 2020 IEEE International Conference on Multimedia and Expo. IEEE (2020)","DOI":"10.1109\/ICME46284.2020.9102868"},{"key":"2237_CR69","doi-asserted-by":"crossref","unstructured":"Zhang, L., Danelljan, M., Gonzalez-Garcia, A., van\u00a0de Weijer, J., Shahbaz\u00a0Khan, F.: Multi-modal fusion for end-to-end rgb-t tracking. In: Proceedings of the IEEE International Conference on Computer Vision Workshops (2019)","DOI":"10.1109\/ICCVW.2019.00278"},{"key":"2237_CR70","doi-asserted-by":"crossref","unstructured":"Zhang, T., Liu, S., Xu, C., Liu, B., Yang, M.H.: Correlation particle filter for visual tracking. IEEE Trans. Image Process. (2017)","DOI":"10.1109\/TIP.2017.2781304"},{"key":"2237_CR71","doi-asserted-by":"publisher","first-page":"365","DOI":"10.1109\/TPAMI.2018.2797062","volume":"41","author":"T Zhang","year":"2018","unstructured":"Zhang, T., Xu, C., Yang, M.H.: Learning multi-task correlation particle filters for visual tracking. IEEE Trans. Pattern Anal. Mach. Intell. 41, 365\u2013378 (2018)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2237_CR72","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Peng, H.: Deeper and wider siamese networks for real-time visual tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2019)","DOI":"10.1109\/CVPR.2019.00472"},{"key":"2237_CR73","doi-asserted-by":"publisher","first-page":"557","DOI":"10.1109\/JETCAS.2020.3040329","volume":"10","author":"Q Zhu","year":"2020","unstructured":"Zhu, Q., Triesch, J., Shi, B.E.: An event-by-event approach for velocity estimation and object tracking with an active event camera. IEEE J. Emerging Sel. Top. Circuits Syst. 10, 557\u2013566 (2020)","journal-title":"IEEE J. Emerging Sel. Top. Circuits Syst."},{"key":"2237_CR74","doi-asserted-by":"crossref","unstructured":"Zhu, Y., Li, C., Luo, B., Tang, J., Wang, X.: Dense feature aggregation and pruning for rgbt tracking. In: Proceedings of the 27th ACM International Conference on Multimedia (2019)","DOI":"10.1145\/3343031.3350928"},{"key":"2237_CR75","doi-asserted-by":"crossref","unstructured":"Zhu, Z., Wang, Q., Li, B., Wu, W., Yan, J., Hu, W.: Distractor-aware siamese networks for visual object tracking. In: Proceedings of the European Conference on Computer Vision (2018)","DOI":"10.1007\/978-3-030-01240-3_7"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-021-02237-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-021-02237-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-021-02237-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,6]],"date-time":"2024-09-06T03:52:58Z","timestamp":1725594778000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-021-02237-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,8,9]]},"references-count":75,"journal-issue":{"issue":"9-11","published-print":{"date-parts":[[2021,9]]}},"alternative-id":["2237"],"URL":"https:\/\/doi.org\/10.1007\/s00371-021-02237-9","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2021,8,9]]},"assertion":[{"value":"10 June 2021","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 August 2021","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Jiqing Zhang, Kai Zhao, Bo Dong, Yingkai Fu, Yuxin Wang, Xin Yang and Baocai Yin declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}