{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T16:33:46Z","timestamp":1769186026850,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":34,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819555666","type":"print"},{"value":"9789819555673","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-5567-3_31","type":"book-chapter","created":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:14:02Z","timestamp":1769116442000},"page":"449-463","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Strengthened Node and\u00a0Edge Generation for\u00a0Enhanced Information Interaction in\u00a0Infrared-Visible Fusion"],"prefix":"10.1007","author":[{"given":"Haohuinan","family":"Zhang","sequence":"first","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,23]]},"reference":[{"key":"31_CR1","doi-asserted-by":"crossref","unstructured":"Cao, B., Sun, Y., Zhu, P., Hu, Q.: Multi-modal gated mixture of local-to-global experts for dynamic image fusion. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 23555\u201323564 (2023)","DOI":"10.1109\/ICCV51070.2023.02153"},{"issue":"3","key":"31_CR2","doi-asserted-by":"crossref","first-page":"3677","DOI":"10.1109\/TPAMI.2022.3174515","volume":"45","author":"C Chen","year":"2023","unstructured":"Chen, C., et al.: Relation matters: foreground-aware graph-based relational reasoning for domain adaptive object detection. IEEE Trans. Pattern Anal. Mach. Intell. 45(3), 3677\u20133694 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"31_CR3","doi-asserted-by":"publisher","first-page":"199","DOI":"10.1016\/j.optcom.2014.12.032","volume":"341","author":"G Cui","year":"2015","unstructured":"Cui, G., Feng, H., Xu, Z., Li, Q., Chen, Y.: Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition. Opt. Commun. 341, 199\u2013209 (2015)","journal-title":"Opt. Commun."},{"issue":"10","key":"31_CR4","doi-asserted-by":"publisher","first-page":"13232","DOI":"10.1109\/TNNLS.2023.3266452","volume":"35","author":"H Fu","year":"2024","unstructured":"Fu, H., et al.: LRAF-Net: long-range attention fusion network for visible\u2013infrared object detection. IEEE Trans. Neural Netw. Learn. Syst. 35(10), 13232\u201313245 (2024)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"31_CR5","doi-asserted-by":"crossref","unstructured":"Fusco, F., Vlachos, M., Vasileiadis, V., Wardatzky, K., Schneider, J.: RecoNet: an interpretable neural architecture for recommender systems. In: Proceedings of the 28th International Joint Conference on Artificial Intelligence, IJCAI 2019, pp. 2343\u20132349. AAAI Press (2019)","DOI":"10.24963\/ijcai.2019\/325"},{"issue":"10","key":"31_CR6","doi-asserted-by":"publisher","first-page":"2913","DOI":"10.1109\/TCSVT.2018.2874312","volume":"29","author":"C Li","year":"2019","unstructured":"Li, C., Zhu, C., Zhang, J., Luo, B., Wu, X., Tang, J.: Learning local-global multi-graph descriptors for RGB-T object tracking. IEEE Trans. Circuits Syst. Video Technol. 29(10), 2913\u20132926 (2019)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"31_CR7","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2024.102839","volume":"117","author":"H Li","year":"2025","unstructured":"Li, H., Ma, H., Cheng, C., Shen, Z., Song, X., Wu, X.J.: Conti-fuse: a novel continuous decomposition-based fusion framework for infrared and visible images. Inf. Fusion 117, 102839 (2025)","journal-title":"Inf. Fusion"},{"key":"31_CR8","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","volume":"73","author":"H Li","year":"2021","unstructured":"Li, H., Wu, X.J., Kittler, J.: RFN-Nest: an end-to-end residual fusion network for infrared and visible images. Inf. Fusion 73, 72\u201386 (2021)","journal-title":"Inf. Fusion"},{"key":"31_CR9","doi-asserted-by":"crossref","unstructured":"Li, J., Chen, J., Liu, J., Ma, H.: Learning a graph neural network with cross modality interaction for image fusion. In: Proceedings of the 31st ACM International Conference on Multimedia, MM 2023, pp. 4471\u20134479. Association for Computing Machinery, New York (2023)","DOI":"10.1145\/3581783.3612135"},{"key":"31_CR10","doi-asserted-by":"crossref","unstructured":"Li, J., Liu, J., Zhou, S., Zhang, Q., Kasabov, N.: Infrared and visible image fusion based on residual dense network and gradient loss. Infrared Phys. Technol. 128, 104486 (2022)","DOI":"10.1016\/j.infrared.2022.104486"},{"key":"31_CR11","doi-asserted-by":"crossref","unstructured":"Liang, P., Jiang, J., Liu, X., Ma, J.: Fusion from decomposition: a self-supervised decomposition approach for image fusion. In: European Conference on Computer Vision (ECCV) (2022)","DOI":"10.1007\/978-3-031-19797-0_41"},{"key":"31_CR12","doi-asserted-by":"crossref","unstructured":"Liu, J., et al.: Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5792\u20135801 (2022)","DOI":"10.1109\/CVPR52688.2022.00571"},{"key":"31_CR13","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., Li, C., Jiang, J.: FusionGAN: a generative adversarial network for infrared and visible image fusion. Inf. Fusion 48, 11\u201326 (2019)","journal-title":"Inf. Fusion"},{"issue":"10","key":"31_CR14","doi-asserted-by":"publisher","first-page":"1360","DOI":"10.1109\/TCSVT.2007.903776","volume":"17","author":"V Maik","year":"2007","unstructured":"Maik, V., Cho, D., Shin, J., Paik, J.: Regularized restoration using image fusion for digital auto-focusing. IEEE Trans. Circuits Syst. Video Technol. 17(10), 1360\u20131369 (2007)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"31_CR15","doi-asserted-by":"crossref","unstructured":"Pei, X., Huang, T., Xu, C.: EfficientVMamba: atrous selective scan for light weight visual mamba. arXiv preprint arXiv:2403.09977 (2024)","DOI":"10.1609\/aaai.v39i6.32690"},{"issue":"9","key":"31_CR16","doi-asserted-by":"publisher","first-page":"5873","DOI":"10.1109\/TPAMI.2024.3368158","volume":"46","author":"M Qiao","year":"2024","unstructured":"Qiao, M., et al.: HyperSOR: context-aware graph hypernetwork for salient object ranking. IEEE Trans. Pattern Anal. Mach. Intell. 46(9), 5873\u20135889 (2024)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"31_CR17","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","volume":"82","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Ma, J.: Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 82, 28\u201342 (2022)","journal-title":"Inf. Fusion"},{"key":"31_CR18","doi-asserted-by":"crossref","unstructured":"Tang, L., Yuan, J., Zhang, H., Jiang, X., Ma, J.: PIAFusion: a progressive infrared and visible image fusion network based on illumination aware. Inf. Fusion (2022)","DOI":"10.1016\/j.inffus.2022.03.007"},{"key":"31_CR19","doi-asserted-by":"publisher","first-page":"5413","DOI":"10.1109\/TMM.2022.3192661","volume":"25","author":"W Tang","year":"2023","unstructured":"Tang, W., He, F., Liu, Y.: YDTR: infrared and visible image fusion via y-shape dynamic transformer. Trans. Multi. 25, 5413\u20135428 (2023)","journal-title":"Trans. Multi."},{"issue":"7","key":"31_CR20","doi-asserted-by":"publisher","first-page":"3159","DOI":"10.1109\/TCSVT.2023.3234340","volume":"33","author":"W Tang","year":"2023","unstructured":"Tang, W., He, F., Liu, Y., Duan, Y., Si, T.: DATFuse: infrared and visible image fusion via dual attention transformer. IEEE Trans. Circuits Syst. Video Technol. 33(7), 3159\u20133172 (2023)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"31_CR21","unstructured":"Team, F.: FREE Teledyne FLIR thermal dataset for algorithm training (2019). https:\/\/www.flir.com\/oem\/adas\/adas-dataset-form\/"},{"key":"31_CR22","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS 2017, pp. 6000\u20136010. Curran Associates Inc., Red Hook (2017)"},{"key":"31_CR23","doi-asserted-by":"crossref","unstructured":"Wang, D., Liu, J., Fan, X., Liu, R.: Unsupervised misaligned infrared and visible image fusion via cross-modality image generation and registration. In: IJCAI, pp. 3508\u20133515 (2022)","DOI":"10.24963\/ijcai.2022\/487"},{"key":"31_CR24","first-page":"1","volume":"71","author":"Z Wang","year":"2022","unstructured":"Wang, Z., Chen, Y., Shao, W., Li, H., Zhang, L.: SwinFuse: a residual SWIN transformer fusion network for infrared and visible images. IEEE Trans. Instrum. Meas. 71, 1\u201312 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"4","key":"31_CR25","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A., Sheikh, H., Simoncelli, E.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"key":"31_CR26","doi-asserted-by":"crossref","unstructured":"Xie, G.S., Liu, J., Xiong, H., Shao, L.: Scale-aware graph neural network for few-shot semantic segmentation. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5471\u20135480 (2021)","DOI":"10.1109\/CVPR46437.2021.00543"},{"issue":"1","key":"31_CR27","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2022","unstructured":"Xu, H., Ma, J., Jiang, J., Guo, X., Ling, H.: U2fusion: a unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 502\u2013518 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"31_CR28","doi-asserted-by":"crossref","unstructured":"Xydeas, C., Petrovic, V.: Objective image fusion performance measure. Electron. Lett. 36, 308\u2013309 (2000)","DOI":"10.1049\/el:20000267"},{"issue":"10","key":"31_CR29","doi-asserted-by":"publisher","first-page":"2761","DOI":"10.1007\/s11263-021-01501-8","volume":"129","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Ma, J.: SDNet: a versatile squeeze-and-decomposition network for real-time image fusion. Int. J. Comput. Vision 129(10), 2761\u20132785 (2021)","journal-title":"Int. J. Comput. Vision"},{"key":"31_CR30","doi-asserted-by":"crossref","unstructured":"Zhang, H., Fromont, \u00c9., Lef\u00e8vre, S., Avignon, B.: Multispectral fusion for object detection with cyclic fuse-and-refine blocks (2020)","DOI":"10.1109\/ICIP40778.2020.9191080"},{"key":"31_CR31","doi-asserted-by":"publisher","first-page":"8988","DOI":"10.1109\/TMM.2023.3243659","volume":"25","author":"J Zhang","year":"2023","unstructured":"Zhang, J., et al.: Transformer based conditional GAN for multimodal image fusion. Trans. Multi. 25, 8988\u20139001 (2023)","journal-title":"Trans. Multi."},{"key":"31_CR32","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","volume":"54","author":"Y Zhang","year":"2020","unstructured":"Zhang, Y., Liu, Y., Sun, P., Yan, H., Zhao, X., Zhang, L.: IFCNN: a general image fusion framework based on convolutional neural network. Inf. Fusion 54, 99\u2013118 (2020)","journal-title":"Inf. Fusion"},{"key":"31_CR33","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Xu, S., Zhang, C., Liu, J., Zhang, J., Li, P.: DIDFuse: deep image decomposition for infrared and visible image fusion. In: Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI 2020 (2021)","DOI":"10.24963\/ijcai.2020\/135"},{"issue":"3","key":"31_CR34","doi-asserted-by":"publisher","first-page":"1186","DOI":"10.1109\/TCSVT.2021.3075745","volume":"32","author":"Z Zhao","year":"2022","unstructured":"Zhao, Z., et al.: Efficient and model-based infrared and visible image fusion via algorithm unrolling. IEEE Trans. Circuits Syst. Video Technol. 32(3), 1186\u20131196 (2022)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-5567-3_31","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:14:07Z","timestamp":1769116447000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-5567-3_31"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819555666","9789819555673"],"references-count":34,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-5567-3_31","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"23 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shanghai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2025.prcv.cn\/index.asp","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}