{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T16:33:30Z","timestamp":1769186010832,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":35,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819555666","type":"print"},{"value":"9789819555673","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-5567-3_21","type":"book-chapter","created":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:13:39Z","timestamp":1769116419000},"page":"298-312","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["PCFusion: A Unified Image Fusion Network with\u00a0Perception-Driven Cross-Domain Learning"],"prefix":"10.1007","author":[{"given":"Siying","family":"Xu","sequence":"first","affiliation":[]},{"given":"Gang","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,23]]},"reference":[{"issue":"4","key":"21_CR1","doi-asserted-by":"publisher","first-page":"2049","DOI":"10.1109\/TIP.2018.2794218","volume":"27","author":"J Cai","year":"2018","unstructured":"Cai, J., Gu, S., Zhang, L.: Learning a deep single image contrast enhancer from multi-exposure images. IEEE Trans. Image Process. 27(4), 2049\u20132062 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"21_CR2","doi-asserted-by":"crossref","unstructured":"Cao, H., et al.: Swin-UNet: UNet-like pure transformer for medical image segmentation. In: European conference on computer vision, pp. 205\u2013218. Springer (2022)","DOI":"10.1007\/978-3-031-25066-8_9"},{"key":"21_CR3","unstructured":"Cao, Z., Zhong, Y., Wang, Z., Deng, L.J.: MMAIF: multi-task and multi-degradation all-in-one for image fusion with language guidance (2025)"},{"key":"21_CR4","doi-asserted-by":"crossref","unstructured":"Chen, L.C., Zhu, Y., Papandreou, G., Schroff, F., Adam, H.: Encoder-decoder with atrous separable convolution for semantic image segmentation. In: Proceedings of the European conference on computer vision (ECCV), pp. 801\u2013818 (2018)","DOI":"10.1007\/978-3-030-01234-2_49"},{"key":"21_CR5","doi-asserted-by":"publisher","first-page":"80","DOI":"10.1016\/j.inffus.2022.11.010","volume":"92","author":"C Cheng","year":"2023","unstructured":"Cheng, C., Xu, T., Wu, X.J.: Mufusion: a general unsupervised image fusion network based on memory unit. Inf. Fusion 92, 80\u201392 (2023)","journal-title":"Inf. Fusion"},{"key":"21_CR6","doi-asserted-by":"publisher","first-page":"3098","DOI":"10.1109\/TIP.2021.3058764","volume":"30","author":"X Deng","year":"2021","unstructured":"Deng, X., Zhang, Y., Xu, M., Gu, S., Duan, Y.: Deep coupled feedback network for joint exposure fusion and image super-resolution. IEEE Trans. Image Process. 30, 3098\u20133112 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"21_CR7","doi-asserted-by":"crossref","unstructured":"Ha, Q., Watanabe, K., Karasawa, T., Ushiku, Y., Harada, T.: MFNet: towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In: 2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5108\u20135115. IEEE (2017)","DOI":"10.1109\/IROS.2017.8206396"},{"key":"21_CR8","unstructured":"Howard, A.G., et al.: Mobilenets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017)"},{"key":"21_CR9","doi-asserted-by":"publisher","first-page":"478","DOI":"10.1016\/j.infrared.2017.07.010","volume":"85","author":"X Jin","year":"2017","unstructured":"Jin, X.: A survey of infrared and visual image fusion methods. Infrared Phys. Technol. 85, 478\u2013501 (2017)","journal-title":"Infrared Phys. Technol."},{"issue":"5","key":"21_CR10","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li, H., Wu, X.J.: DenseFuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2018)","journal-title":"IEEE Trans. Image Process."},{"issue":"12","key":"21_CR11","doi-asserted-by":"publisher","first-page":"9645","DOI":"10.1109\/TIM.2020.3005230","volume":"69","author":"H Li","year":"2020","unstructured":"Li, H., Wu, X.J., Durrani, T.: NestFuse: an infrared and visible image fusion architecture based on nest connection and spatial\/channel attention models. IEEE Trans. Instrum. Meas. 69(12), 9645\u20139656 (2020)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"21_CR12","doi-asserted-by":"crossref","unstructured":"Liu, Z., et\u00a0al.: Swin transformer v2: scaling up capacity and resolution. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 12009\u201312019 (2022)","DOI":"10.1109\/CVPR52688.2022.01170"},{"key":"21_CR13","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"21_CR14","doi-asserted-by":"publisher","unstructured":"Liu, Z., et al.: Bi-level dynamic learning for jointly multi-modality image fusion and beyond. In: Elkind, E. (ed.) Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, IJCAI-23, pp. 1240\u20131248. International Joint Conferences on Artificial Intelligence Organization (2023). https:\/\/doi.org\/10.24963\/ijcai.2023\/138, main Track","DOI":"10.24963\/ijcai.2023\/138"},{"key":"21_CR15","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1016\/j.inffus.2018.02.004","volume":"45","author":"J Ma","year":"2019","unstructured":"Ma, J., Ma, Y., Li, C.: Infrared and visible image fusion methods and applications: a survey. Inf. Fusion 45, 153\u2013178 (2019)","journal-title":"Inf. Fusion"},{"issue":"7","key":"21_CR16","doi-asserted-by":"publisher","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","volume":"9","author":"J Ma","year":"2022","unstructured":"Ma, J., et al.: SwinFusion: cross-domain long-range learning for general image fusion via Swin transformer. IEEE\/CAA J. Automatica Sinica 9(7), 1200\u20131217 (2022)","journal-title":"IEEE\/CAA J. Automatica Sinica"},{"key":"21_CR17","unstructured":"Ma, L., et al.: Causality-driven infrared and visible image fusion (2025)"},{"key":"21_CR18","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2014.10.004","volume":"25","author":"M Nejati","year":"2015","unstructured":"Nejati, M., Samavi, S., Shirani, S.: Multi-focus image fusion using dictionary-based sparse representation. Inf. Fusion 25, 72\u201384 (2015)","journal-title":"Inf. Fusion"},{"key":"21_CR19","doi-asserted-by":"crossref","unstructured":"Qu, L., Liu, S., Wang, M., Song, Z.: Transmef: a transformer-based multi-exposure image fusion framework using self-supervised multi-task learning. In: Proceedings of the AAAI conference on artificial intelligence, vol.\u00a036, pp. 2126\u20132134 (2022)","DOI":"10.1609\/aaai.v36i2.20109"},{"key":"21_CR20","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 779\u2013788 (2016)","DOI":"10.1109\/CVPR.2016.91"},{"key":"21_CR21","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","volume":"82","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Ma, J.: Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 82, 28\u201342 (2022)","journal-title":"Inf. Fusion"},{"key":"21_CR22","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2022.109295","volume":"137","author":"W Tang","year":"2023","unstructured":"Tang, W., He, F., Liu, Y.: TCCFusion: an infrared and visible image fusion method based on transformer and cross correlation. Pattern Recogn. 137, 109295 (2023)","journal-title":"Pattern Recogn."},{"key":"21_CR23","doi-asserted-by":"publisher","unstructured":"Wang, K., et al.: Infrared and visible image fusion driven by multimodal large language models. Front. Physics 13 (2025). https:\/\/doi.org\/10.3389\/fphy.2025.1599937, https:\/\/www.frontiersin.org\/articles\/10.3389\/fphy.2025.1599937","DOI":"10.3389\/fphy.2025.1599937"},{"issue":"4","key":"21_CR24","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"key":"21_CR25","doi-asserted-by":"crossref","unstructured":"Wu, G., et al.: Every SAM drop counts: embracing semantic priors for multi-modality image fusion and beyond (2025)","DOI":"10.1109\/CVPR52734.2025.01666"},{"key":"21_CR26","doi-asserted-by":"crossref","unstructured":"Wu, Y., Liu, Z., Liu, J., Fan, X., Liu, R.: Breaking free from fusion rule: a fully semantic-driven infrared and visible image fusion. IEEE Sig. Process. Lett. (2023)","DOI":"10.1109\/LSP.2023.3266980"},{"issue":"1","key":"21_CR27","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2020","unstructured":"Xu, H., Ma, J., Jiang, J., Guo, X., Ling, H.: U2Fusion: a unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 502\u2013518 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"21_CR28","doi-asserted-by":"publisher","first-page":"7203","DOI":"10.1109\/TIP.2020.2999855","volume":"29","author":"H Xu","year":"2020","unstructured":"Xu, H., Ma, J., Zhang, X.P.: MEF-GAN: multi-exposure image fusion via generative adversarial networks. IEEE Trans. Image Process. 29, 7203\u20137216 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"21_CR29","doi-asserted-by":"publisher","first-page":"1521603","DOI":"10.3389\/fnbot.2024.1521603","volume":"18","author":"Z Yang","year":"2024","unstructured":"Yang, Z., Li, Y., Tang, X., Xie, M.: MGFusion: a multimodal large language model-guided information perception for infrared and visible image fusion. Front. Neurorobot. 18, 1521603 (2024). https:\/\/doi.org\/10.3389\/fnbot.2024.1521603","journal-title":"Front. Neurorobot."},{"key":"21_CR30","doi-asserted-by":"crossref","unstructured":"Zeiler, M.D., Taylor, G.W., Fergus, R.: Adaptive deconvolutional networks for mid and high level feature learning. In: 2011 international conference on computer vision, pp. 2018\u20132025. IEEE (2011)","DOI":"10.1109\/ICCV.2011.6126474"},{"key":"21_CR31","doi-asserted-by":"publisher","first-page":"40","DOI":"10.1016\/j.inffus.2020.08.022","volume":"66","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Le, Z., Shao, Z., Xu, H., Ma, J.: MFF-GAN: an unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Inf. Fusion 66, 40\u201353 (2021)","journal-title":"Inf. Fusion"},{"key":"21_CR32","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1016\/j.inffus.2021.06.008","volume":"76","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Xu, H., Tian, X., Jiang, J., Ma, J.: Image fusion meets deep learning: a survey and perspective. Inf. Fusion 76, 323\u2013336 (2021). https:\/\/doi.org\/10.1016\/j.inffus.2021.06.008","journal-title":"Inf. Fusion"},{"key":"21_CR33","doi-asserted-by":"publisher","first-page":"111","DOI":"10.1016\/j.inffus.2021.02.005","volume":"74","author":"X Zhang","year":"2021","unstructured":"Zhang, X.: Benchmarking and comparing multi-exposure image fusion algorithms. Inf. Fusion 74, 111\u2013131 (2021)","journal-title":"Inf. Fusion"},{"key":"21_CR34","doi-asserted-by":"crossref","unstructured":"Zhao, Z., et al.: CDDFuse: correlation-driven dual-branch feature decomposition for multi-modality image fusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5906\u20135916 (2023)","DOI":"10.1109\/CVPR52729.2023.00572"},{"key":"21_CR35","doi-asserted-by":"crossref","unstructured":"Zhao, Z., et al.: Equivariant multi-modality image fusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 25912\u201325921 (2024)","DOI":"10.1109\/CVPR52733.2024.02448"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-5567-3_21","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:13:43Z","timestamp":1769116423000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-5567-3_21"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819555666","9789819555673"],"references-count":35,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-5567-3_21","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"23 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shanghai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2025.prcv.cn\/index.asp","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}