{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T17:36:11Z","timestamp":1772645771277,"version":"3.50.1"},"reference-count":57,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"The Postgraduate Innovation Fund Project Southwest University of Science and Technology","award":["24ycx1061"],"award-info":[{"award-number":["24ycx1061"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2026,1]]},"DOI":"10.1007\/s00371-025-04310-z","type":"journal-article","created":{"date-parts":[[2026,1,25]],"date-time":"2026-01-25T10:37:18Z","timestamp":1769337438000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["DDCFusion: enhancing visible\u2013infrared image fusion via dual-domain collaborative learning"],"prefix":"10.1007","volume":"42","author":[{"given":"HaoXiang","family":"Weng","sequence":"first","affiliation":[]},{"given":"Bo","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Linjing","family":"Li","sequence":"additional","affiliation":[]},{"given":"Hua","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Kaiming","family":"Cao","sequence":"additional","affiliation":[]},{"given":"Xueci","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Hongrui","family":"Miao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,25]]},"reference":[{"key":"4310_CR1","doi-asserted-by":"publisher","first-page":"100","DOI":"10.1016\/j.inffus.2016.05.004","volume":"33","author":"S Li","year":"2017","unstructured":"Li, S., Kang, X., Fang, L., et al.: Pixel-level image fusion: a survey of the state of the art. Inf. Fusion 33, 100\u2013112 (2017)","journal-title":"Inf. Fusion"},{"issue":"9","key":"4310_CR2","first-page":"4819","volume":"44","author":"X Zhang","year":"2021","unstructured":"Zhang, X.: Deep learning-based multi-focus image fusion: a survey and a comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 44(9), 4819\u20134838 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4310_CR3","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1016\/j.inffus.2022.09.019","volume":"90","author":"S Karim","year":"2023","unstructured":"Karim, S., Tong, G., Li, J., et al.: Current advances and future perspectives of image fusion: a comprehensive review. Inf. Fusion 90, 185\u2013217 (2023)","journal-title":"Inf. Fusion"},{"key":"4310_CR4","doi-asserted-by":"publisher","first-page":"2226","DOI":"10.1109\/TMM.2022.3144890","volume":"25","author":"N Jiang","year":"2022","unstructured":"Jiang, N., Sheng, B., Li, P., et al.: Photohelper: portrait photographing guidance via deep feature retrieval and fusion. IEEE Trans. Multimedia 25, 2226\u20132238 (2022)","journal-title":"IEEE Trans. Multimedia"},{"key":"4310_CR5","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1016\/j.inffus.2018.02.004","volume":"45","author":"J Ma","year":"2019","unstructured":"Ma, J., Ma, Y., Li, C.: Infrared and visible image fusion methods and applications: a survey. Inf. Fusion 45, 153\u2013178 (2019)","journal-title":"Inf. Fusion"},{"issue":"7","key":"4310_CR6","doi-asserted-by":"publisher","first-page":"2864","DOI":"10.1109\/TIP.2013.2244222","volume":"22","author":"S Li","year":"2013","unstructured":"Li, S., Kang, X., Hu, J.: Image fusion with guided filtering. IEEE Trans. Image Process. 22(7), 2864\u20132875 (2013)","journal-title":"IEEE Trans. Image Process."},{"issue":"10","key":"4310_CR7","doi-asserted-by":"publisher","first-page":"7719","DOI":"10.1109\/TNNLS.2022.3146004","volume":"34","author":"Z Yu","year":"2023","unstructured":"Yu, Z., Zhihua, C., Ping, L., Haitao, S., et al.: FSAD-net: feedback spatial attention dehazing network. IEEE Trans Neural Netw Learni Syst 34(10), 7719\u20137733 (2023)","journal-title":"IEEE Trans Neural Netw Learni Syst"},{"key":"4310_CR8","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1109\/TMM.2021.3120873","volume":"25","author":"X Lin","year":"2021","unstructured":"Lin, X., Sun, S., Huang, W., et al.: EAPT: efficient attention pyramid transformer for image processing. IEEE Trans. Multimedia 25, 50\u201361 (2021)","journal-title":"IEEE Trans. Multimedia"},{"key":"4310_CR9","doi-asserted-by":"publisher","first-page":"161","DOI":"10.1016\/j.patcog.2018.08.005","volume":"85","author":"C Li","year":"2019","unstructured":"Li, C., Song, D., Tong, R., et al.: Illumination-aware faster R-CNN for robust multispectral pedestrian detection. Pattern Recognit. 85, 161\u2013171 (2019)","journal-title":"Pattern Recognit."},{"key":"4310_CR10","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1016\/j.inffus.2021.06.008","volume":"76","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Xu, H., Tian, X., et al.: Image fusion meets deep learning: a survey and perspective. Inf. Fusion 76, 323\u2013336 (2021)","journal-title":"Inf. Fusion"},{"issue":"12","key":"4310_CR11","doi-asserted-by":"publisher","first-page":"1882","DOI":"10.1109\/LSP.2016.2618776","volume":"23","author":"Y Liu","year":"2016","unstructured":"Liu, Y., Chen, X., Ward, R.K., et al.: Image fusion with convolutional sparse representation. IEEE Signal Process. Lett. 23(12), 1882\u20131886 (2016)","journal-title":"IEEE Signal Process. Lett."},{"issue":"5","key":"4310_CR12","doi-asserted-by":"publisher","first-page":"1074","DOI":"10.1109\/JSTSP.2011.2112332","volume":"5","author":"N Yu","year":"2011","unstructured":"Yu, N., Qiu, T., Bi, F., et al.: Image features extraction and fusion based on joint sparse representation. IEEE J. Sel. Top. Signal Process. 5(5), 1074\u20131082 (2011)","journal-title":"IEEE J. Sel. Top. Signal Process."},{"key":"4310_CR13","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1016\/j.ins.2019.08.066","volume":"508","author":"J Chen","year":"2020","unstructured":"Chen, J., Li, X., Luo, L., et al.: Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf. Sci. 508, 64\u201378 (2020)","journal-title":"Inf. Sci."},{"key":"4310_CR14","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.infrared.2015.11.002","volume":"74","author":"H Li","year":"2016","unstructured":"Li, H., Liu, L., Huang, W., et al.: An improved fusion algorithm for infrared and visible images based on multi-scale transform. Infrared Phys. Technol. 74, 28\u201337 (2016)","journal-title":"Infrared Phys. Technol."},{"issue":"5","key":"4310_CR15","doi-asserted-by":"publisher","first-page":"743","DOI":"10.1109\/JSEN.2007.894926","volume":"7","author":"N Cvejic","year":"2007","unstructured":"Cvejic, N., Bull, D., Canagarajah, N.: Region-based multimodal image fusion using ICA bases. IEEE Sens. J. 7(5), 743\u2013751 (2007)","journal-title":"IEEE Sens. J."},{"key":"4310_CR16","first-page":"1046","volume":"2","author":"J Mou","year":"2013","unstructured":"Mou, J., Gao, W., Song, Z.: Image fusion based on non-negative matrix factorization and infrared feature extraction[C]\/\/2013 6th international congress on image and signal processing (CISP). IEEE 2, 1046\u20131050 (2013)","journal-title":"IEEE"},{"key":"4310_CR17","doi-asserted-by":"publisher","first-page":"52","DOI":"10.1016\/j.infrared.2016.01.009","volume":"76","author":"DP Bavirisetti","year":"2016","unstructured":"Bavirisetti, D.P., Dhuli, R.: Two-scale image fusion of visible and infrared images using saliency detection. Infrared Phys. Technol. 76, 52\u201364 (2016)","journal-title":"Infrared Phys. Technol."},{"key":"4310_CR18","doi-asserted-by":"publisher","first-page":"94","DOI":"10.1016\/j.infrared.2017.04.018","volume":"83","author":"CH Liu","year":"2017","unstructured":"Liu, C.H., Qi, Y., Ding, W.R.: Infrared and visible image fusion method based on saliency detection in sparse domain. Infrared Phys. Technol. 83, 94\u2013102 (2017)","journal-title":"Infrared Phys. Technol."},{"key":"4310_CR19","doi-asserted-by":"publisher","first-page":"162","DOI":"10.1109\/TCI.2022.3151472","volume":"8","author":"H Yan","year":"2022","unstructured":"Yan, H., Zhang, J.X., Zhang, X.: Injected infrared and visible image fusion via decomposition model and guided filtering. IEEE Trans Comput. Imaging 8, 162\u2013173 (2022)","journal-title":"IEEE Trans Comput. Imaging"},{"key":"4310_CR20","doi-asserted-by":"publisher","first-page":"158","DOI":"10.1016\/j.inffus.2017.10.007","volume":"42","author":"Y Liu","year":"2018","unstructured":"Liu, Y., Chen, X., Wang, Z., et al.: Deep learning for pixel-level image fusion: recent advances and future prospects. Inf. Fusion 42, 158\u2013173 (2018)","journal-title":"Inf. Fusion"},{"key":"4310_CR21","doi-asserted-by":"publisher","DOI":"10.1109\/TTE.2025.3562604","author":"H Yang","year":"2025","unstructured":"Yang, H., Liu, Z., Ma, N., et al.: CSRM-MIM: a self-supervised pre-training method for detecting catenary support components in electrified railways. IEEE Trans. Transp. Electrif. (2025). https:\/\/doi.org\/10.1109\/TTE.2025.3562604","journal-title":"IEEE Trans. Transp. Electrif."},{"key":"4310_CR22","doi-asserted-by":"crossref","unstructured":"Meng, D., Li, S., Wu, H., et al. Semi-supervised multi-modal medical image segmentation for complex situations. arXiv preprint arXiv:2506.17136, 2025.","DOI":"10.1007\/978-3-032-04984-1_47"},{"issue":"26","key":"4310_CR23","doi-asserted-by":"publisher","first-page":"43838","DOI":"10.1364\/OE.504606","volume":"31","author":"E Xypakis","year":"2023","unstructured":"Xypakis, E., De Turris, V., Gala, F., et al.: Physics-informed deep neural network for image denoising. Opt. Express 31(26), 43838\u201343849 (2023)","journal-title":"Opt. Express"},{"key":"4310_CR24","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","volume":"54","author":"Y Zhang","year":"2020","unstructured":"Zhang, Y., Liu, Y., Sun, P., et al.: IFCNN: a general image fusion framework based on convolutional neural network. Inf. Fusion 54, 99\u2013118 (2020)","journal-title":"Inf. Fusion"},{"key":"4310_CR25","first-page":"1","volume":"71","author":"Z Wang","year":"2022","unstructured":"Wang, Z., Chen, Y., Shao, W., et al.: SwinFuse: a residual swin transformer fusion network for infrared and visible images. IEEE Trans. Instrum. Meas. 71, 1\u201312 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"4310_CR26","unstructured":"Vaswani A, Shazeer N, Parmar N, et al. Attention is all you need[J]. Advances in neural information processing systems, 2017, 30."},{"key":"4310_CR27","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Bai, H., Zhang, J., et al. Cddfuse: Correlation-driven dual-branch feature decomposition for multi-modality image fusion. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. 2023: 5906\u20135916.","DOI":"10.1109\/CVPR52729.2023.00572"},{"key":"4310_CR28","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.101870","volume":"99","author":"L Tang","year":"2023","unstructured":"Tang, L., Zhang, H., Xu, H., et al.: Rethinking the necessity of image fusion in high-level vision tasks: a practical infrared and visible image fusion network based on progressive semantic injection and scene fidelity. Inf. Fusion 99, 101870 (2023)","journal-title":"Inf. Fusion"},{"issue":"11\u201326","key":"4310_CR29","first-page":"20","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., et al.: FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. fusion 48(11\u201326), 20 (2019)","journal-title":"Inf. fusion"},{"key":"4310_CR30","doi-asserted-by":"publisher","first-page":"336","DOI":"10.1016\/j.inffus.2022.12.007","volume":"92","author":"Y Rao","year":"2023","unstructured":"Rao, Y., Wu, D., Han, M., et al.: AT-GAN: a generative adversarial network with attention and transition for infrared and visible image fusion. Inf. Fusion 92, 336\u2013349 (2023)","journal-title":"Inf. Fusion"},{"key":"4310_CR31","first-page":"1","volume":"70","author":"H Xu","year":"2021","unstructured":"Xu, H., Wang, X., Ma, J.: DRF: disentangled representation for visible and infrared image fusion. IEEE Trans. Instrum. Meas. 70, 1\u201313 (2021)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"4310_CR32","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., et al.: FusionGAN: a generative adversarial network for infrared and visible image fusion. Inf. Fusion 48, 11\u201326 (2019)","journal-title":"Inf. Fusion"},{"key":"4310_CR33","doi-asserted-by":"publisher","first-page":"4980","DOI":"10.1109\/TIP.2020.2977573","volume":"29","author":"J Ma","year":"2020","unstructured":"Ma, J., Xu, H., Jiang, J., et al.: DDcGAN: A dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans. Image Process. 29, 4980\u20134995 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"4310_CR34","doi-asserted-by":"publisher","first-page":"1383","DOI":"10.1109\/TMM.2020.2997127","volume":"23","author":"J Li","year":"2020","unstructured":"Li, J., Huo, H., Li, C., et al.: AttentionFGAN: infrared and visible image fusion using attention-based generative adversarial networks. IEEE Trans. Multimedia 23, 1383\u20131396 (2020)","journal-title":"IEEE Trans. Multimedia"},{"issue":"1","key":"4310_CR35","doi-asserted-by":"publisher","DOI":"10.1002\/cav.2201","volume":"35","author":"X Zhu","year":"2024","unstructured":"Zhu, X., Yao, X., Zhang, J., et al.: TMSDNet: transformer with multi\u2010scale dense network for single and multi\u2010view 3D reconstruction. Comput. Anim. Virtual Worlds 35(1), e2201 (2024)","journal-title":"Comput. Anim. Virtual Worlds"},{"issue":"1","key":"4310_CR36","doi-asserted-by":"publisher","DOI":"10.1002\/cav.2203","volume":"35","author":"Y Liu","year":"2024","unstructured":"Liu, Y., Huang, E., Zhou, Z., et al.: 3D facial attractiveness prediction based on deep feature fusion. Comput. Anim. Virtual Worlds 35(1), e2203 (2024)","journal-title":"Comput. Anim. Virtual Worlds"},{"key":"4310_CR37","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2022.116905","volume":"200","author":"X Yang","year":"2022","unstructured":"Yang, X., Huo, H., Li, J., et al.: DSG-fusion: infrared and visible image fusion via generative adversarial networks and guided filter. Expert Syst. Appl. 200, 116905 (2022)","journal-title":"Expert Syst. Appl."},{"issue":"9","key":"4310_CR38","doi-asserted-by":"publisher","first-page":"1855","DOI":"10.1016\/j.patcog.2004.03.010","volume":"37","author":"G Pajares","year":"2004","unstructured":"Pajares, G., De La Cruz, J.M.: A wavelet-based image fusion tutorial. Pattern Recognit. 37(9), 1855\u20131872 (2004)","journal-title":"Pattern Recognit."},{"issue":"4","key":"4310_CR39","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1016\/j.isprsjprs.2007.05.009","volume":"62","author":"K Amolins","year":"2007","unstructured":"Amolins, K., Zhang, Y., Dare, P.: Wavelet based image fusion techniques\u2014an introduction, review and comparison. ISPRS J. Photogramm. Remote Sens. 62(4), 249\u2013263 (2007)","journal-title":"ISPRS J. Photogramm. Remote Sens."},{"key":"4310_CR40","doi-asserted-by":"publisher","DOI":"10.1109\/TIM.2024.3370752","author":"H Chen","year":"2024","unstructured":"Chen, H., Deng, L., Chen, Z., et al.: SFCFusion: spatial-frequency collaborative infrared and visible image fusion. IEEE Trans. Instrum. Meas. (2024). https:\/\/doi.org\/10.1109\/TIM.2024.3370752","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"4310_CR41","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2024.108013","volume":"133","author":"C Wang","year":"2024","unstructured":"Wang, C., Wu, J., Fang, A., et al.: An efficient frequency domain fusion network of infrared and visible images. Eng. Appl. Artif. Intell. 133, 108013 (2024)","journal-title":"Eng. Appl. Artif. Intell."},{"key":"4310_CR42","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2024.102359","volume":"109","author":"N Zheng","year":"2024","unstructured":"Zheng, N., Zhou, M., Huang, J., et al.: Frequency integration and spatial compensation network for infrared and visible image fusion. Inf. Fusion 109, 102359 (2024)","journal-title":"Inf. Fusion"},{"key":"4310_CR43","doi-asserted-by":"crossref","unstructured":"Burt, P.J., Adelson, E.H. The Laplacian pyramid as a compact image code. Readings in computer vision. Morgan Kaufmann, 1987: 671\u2013679.","DOI":"10.1016\/B978-0-08-051581-6.50065-9"},{"issue":"4","key":"4310_CR44","doi-asserted-by":"publisher","first-page":"723","DOI":"10.1137\/0515056","volume":"15","author":"A Grossmann","year":"1984","unstructured":"Grossmann, A., Morlet, J.: Decomposition of Hardy functions into square integrable wavelets of constant shape. SIAM J. Math. Anal. 15(4), 723\u2013736 (1984)","journal-title":"SIAM J. Math. Anal."},{"issue":"4","key":"4310_CR45","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., et al.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"key":"4310_CR46","doi-asserted-by":"crossref","unstructured":"Jia, X., Zhu, C., Li, M., et al. LLVIP: A visible-infrared paired dataset for low-light vision. In: Proceedings of the IEEE\/CVF international conference on computer vision. 2021: 3496\u20133504.","DOI":"10.1109\/ICCVW54120.2021.00389"},{"key":"4310_CR47","doi-asserted-by":"crossref","unstructured":"Liu, J., Fan, X., Huang, Z., et al. Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. 2022: 5802\u20135811.","DOI":"10.1109\/CVPR52688.2022.00571"},{"key":"4310_CR48","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1016\/j.inffus.2022.03.007","volume":"83","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Zhang, H., et al.: PIAFusion: A progressive infrared and visible image fusion network based on illumination aware. Inf. Fusion 83, 79\u201392 (2022)","journal-title":"Inf. Fusion"},{"key":"4310_CR49","doi-asserted-by":"publisher","DOI":"10.1016\/j.dsp.2023.104020","volume":"137","author":"S Singh","year":"2023","unstructured":"Singh, S., Singh, H., Bueno, G., et al.: A review of image fusion: methods, applications and performance metrics. Digit. Signal Process. 137, 104020 (2023)","journal-title":"Digit. Signal Process."},{"key":"4310_CR50","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","volume":"82","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Ma, J.: Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 82, 28\u201342 (2022)","journal-title":"Inf. Fusion"},{"key":"4310_CR51","doi-asserted-by":"crossref","unstructured":"Wang, D., Liu, J., Fan, X., et al. Unsupervised misaligned infrared and visible image fusion via cross-modality image generation and registration. arXiv preprint arXiv:2205.11876, 2022.","DOI":"10.24963\/ijcai.2022\/487"},{"key":"4310_CR52","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2024.3412743","author":"D Wang","year":"2024","unstructured":"Wang, D., Liu, J., Ma, L., et al.: Improving misaligned multi-modality image fusion with one-stage progressive dense registration. IEEE Trans. Circuits Syst. Video Technol. (2024). https:\/\/doi.org\/10.1109\/TCSVT.2024.3412743","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"4310_CR53","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2024.102352","volume":"108","author":"X Liu","year":"2024","unstructured":"Liu, X., Huo, H., Li, J., et al.: A semantic-driven coupled network for infrared and visible image fusion. Inf. Fusion 108, 102352 (2024)","journal-title":"Inf. Fusion"},{"key":"4310_CR54","first-page":"482","volume":"2024","author":"K Hu","year":"2024","unstructured":"Hu, K., Zhang, Q., Yuan, M., et al.: SFDFusion: an efficient spatial-frequency domain fusion network for infrared and visible image fusion. IOS Press 2024, 482\u2013489 (2024)","journal-title":"IOS Press"},{"key":"4310_CR55","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Bochkovskiy, A., Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. 2023: 7464\u20137475.","DOI":"10.1109\/CVPR52729.2023.00721"},{"key":"4310_CR56","doi-asserted-by":"crossref","unstructured":"Yujie, L., Xiaorui, S., Wenbin, S., Yafu, Y., et al. S2ANet: combining local spectral and spatial point grouping for point cloud processing, cnki, 2024: 267\u2013279.","DOI":"10.1016\/j.vrih.2023.06.005"},{"key":"4310_CR57","doi-asserted-by":"publisher","DOI":"10.1016\/j.ast.2025.110371","author":"L Li","year":"2025","unstructured":"Li, L., Tian, Y., Deng, X., et al.: Deep reconstruction of Schlieren images for scramjet combustion flow field analysis using FCSD-Net with feature completion strategy. Aerosp. Sci. Technol. (2025). https:\/\/doi.org\/10.1016\/j.ast.2025.110371","journal-title":"Aerosp. Sci. Technol."}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04310-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04310-z","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04310-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T14:02:26Z","timestamp":1772632946000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04310-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1]]},"references-count":57,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2026,1]]}},"alternative-id":["4310"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04310-z","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1]]},"assertion":[{"value":"27 May 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 December 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 January 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"140"}}