{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T04:53:42Z","timestamp":1743051222112,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":33,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819786848"},{"type":"electronic","value":"9789819786855"}],"license":[{"start":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T00:00:00Z","timestamp":1730592000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T00:00:00Z","timestamp":1730592000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-97-8685-5_17","type":"book-chapter","created":{"date-parts":[[2024,11,2]],"date-time":"2024-11-02T05:11:50Z","timestamp":1730524310000},"page":"238-251","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["MIAFusion: Infrared and Visible Image Fusion via Multi-scale Spatial and Channel-Aware Interaction Attention"],"prefix":"10.1007","author":[{"given":"Teng","family":"Lin","sequence":"first","affiliation":[]},{"given":"Ming","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Min","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Jun","family":"Kong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,3]]},"reference":[{"key":"17_CR1","unstructured":"Toet, A.: TNO Image Fusion Dataset (2014). https:\/\/figshare.com\/articles\/dataset\/TNO_Image_Fusion_Dataset\/1008029"},{"key":"17_CR2","doi-asserted-by":"crossref","unstructured":"Chen, C.R., Fan, Q., Panda, R.: CrossViT: cross-attention multi-scale vision transformer for image classification. In: 2021 IEEE\/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, 10\u201317 Oct 2021, pp. 347\u2013356. IEEE (2021)","DOI":"10.1109\/ICCV48922.2021.00041"},{"key":"17_CR3","doi-asserted-by":"publisher","first-page":"199","DOI":"10.1016\/j.optcom.2014.12.032","volume":"341","author":"G Cui","year":"2015","unstructured":"Cui, G., Feng, H., Xu, Z., Li, Q., Chen, Y.: Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition. Opt. Commun. 341, 199\u2013209 (2015)","journal-title":"Opt. Commun."},{"key":"17_CR4","doi-asserted-by":"crossref","unstructured":"d\u2019Ascoli, S., Touvron, H., Leavitt, M.L., Morcos, A.S., Biroli, G., Sagun, L.: ConViT: improving vision transformers with soft convolutional inductive biases. J. Stat. Mech. Theory Exp. 2022(11) (2022)","DOI":"10.1088\/1742-5468\/ac9830"},{"issue":"12","key":"17_CR5","doi-asserted-by":"publisher","first-page":"2959","DOI":"10.1109\/26.477498","volume":"43","author":"AM Eskicioglu","year":"1995","unstructured":"Eskicioglu, A.M., Fisher, P.S.: Image quality measures and their performance. IEEE Trans. Commun. 43(12), 2959\u20132965 (1995)","journal-title":"IEEE Trans. Commun."},{"key":"17_CR6","unstructured":"Fan, Q., Huang, H., Guan, J., He, R.: Rethinking local perception in lightweight vision transformer (2023). CoRR arXiv:abs\/2303.17803"},{"issue":"2","key":"17_CR7","doi-asserted-by":"publisher","first-page":"127","DOI":"10.1016\/j.inffus.2011.08.002","volume":"14","author":"Y Han","year":"2013","unstructured":"Han, Y., Cai, Y., Cao, Y., Xu, X.: A new image fusion performance metric based on visual information fidelity. Inf. Fusion 14(2), 127\u2013135 (2013)","journal-title":"Inf. Fusion"},{"key":"17_CR8","doi-asserted-by":"crossref","unstructured":"Lei, T., Sun, R., Wang, X., Wang, Y., He, X., Nandi, A.K.: CiT-Net: convolutional neural networks hand in hand with vision transformers for medical image segmentation. In: Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, IJCAI 2023, 19th\u201325th Aug 2023, Macao, SAR, China, pp. 1017\u20131025 (2023)","DOI":"10.24963\/ijcai.2023\/113"},{"issue":"5","key":"17_CR9","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2019","unstructured":"Li, H., Wu, X.: DenseFuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2019)","journal-title":"IEEE Trans. Image Process."},{"issue":"12","key":"17_CR10","doi-asserted-by":"publisher","first-page":"9645","DOI":"10.1109\/TIM.2020.3005230","volume":"69","author":"H Li","year":"2020","unstructured":"Li, H., Wu, X., Durrani, T.S.: NestFuse: an infrared and visible image fusion architecture based on nest connection and spatial\/channel attention models. IEEE Trans. Instrum. Meas. 69(12), 9645\u20139656 (2020)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"17_CR11","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","volume":"73","author":"H Li","year":"2021","unstructured":"Li, H., Wu, X., Kittler, J.: RFN-Nest: an end-to-end residual fusion network for infrared and visible images. Inf. Fusion 73, 72\u201386 (2021)","journal-title":"Inf. Fusion"},{"key":"17_CR12","doi-asserted-by":"publisher","first-page":"1383","DOI":"10.1109\/TMM.2020.2997127","volume":"23","author":"J Li","year":"2021","unstructured":"Li, J., Huo, H., Li, C., Wang, R., Feng, Q.: AttentionFGAN: infrared and visible image fusion using attention-based generative adversarial networks. IEEE Trans. Multim. 23, 1383\u20131396 (2021)","journal-title":"IEEE Trans. Multim."},{"key":"17_CR13","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2022.3218574","volume":"71","author":"J Li","year":"2022","unstructured":"Li, J., Zhu, J., Li, C., Chen, X., Yang, B.: CGTF: convolution-guided transformer for infrared and visible image fusion. IEEE Trans. Instrum. Meas. 71, 1\u201314 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"17_CR14","doi-asserted-by":"crossref","unstructured":"Liu, J., et al.: Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, 18\u201324 June 2022, pp. 5792\u20135801 (2022)","DOI":"10.1109\/CVPR52688.2022.00571"},{"issue":"13","key":"17_CR15","doi-asserted-by":"publisher","first-page":"3233","DOI":"10.3390\/rs14133233","volume":"14","author":"X Liu","year":"2022","unstructured":"Liu, X., Gao, H., Miao, Q., Xi, Y., Ai, Y., Gao, D.: MFST: multi-modal feature self-adaptive transformer for infrared and visible image fusion. Remote. Sens. 14(13), 3233 (2022)","journal-title":"Remote. Sens."},{"key":"17_CR16","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: 2021 IEEE\/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, 10\u201317 Oct 2021, pp. 9992\u201310002 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"17_CR17","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2023.3329222","volume":"72","author":"M Lu","year":"2023","unstructured":"Lu, M., Jiang, M., Kong, J., Tao, X.: LDRepFM: a real-time end-to-end visible and infrared image fusion model based on layer decomposition and re-parameterization. IEEE Trans. Instrum. Meas. 72, 1\u201312 (2023)","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"7","key":"17_CR18","doi-asserted-by":"publisher","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","volume":"9","author":"J Ma","year":"2022","unstructured":"Ma, J., Tang, L., Fan, F., Huang, J., Mei, X., Ma, Y.: SwinFusion: cross-domain long-range learning for general image fusion via Swin transformer. IEEE CAA J. Autom. Sinica 9(7), 1200\u20131217 (2022)","journal-title":"IEEE CAA J. Autom. Sinica"},{"issue":"7","key":"17_CR19","doi-asserted-by":"publisher","first-page":"313","DOI":"10.1049\/el:20020212","volume":"38","author":"G Qu","year":"2002","unstructured":"Qu, G., Zhang, D., Yan, P.: Information measure for performance of image fusion. Electron. Lett. 38(7), 313\u2013315 (2002)","journal-title":"Electron. Lett."},{"key":"17_CR20","doi-asserted-by":"crossref","unstructured":"Roberts, J.W., van Aardt, J., Ahmed, F.: Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J. Appl. Remote Sens. 2 (2008)","DOI":"10.1117\/1.2945910"},{"issue":"3","key":"17_CR21","doi-asserted-by":"publisher","first-page":"2576","DOI":"10.1109\/LRA.2019.2904733","volume":"4","author":"Y Sun","year":"2019","unstructured":"Sun, Y., Zuo, W., Liu, M.: RTFNet: rgb-thermal fusion network for semantic segmentation of urban scenes. IEEE Robot. Autom. Lett. 4(3), 2576\u20132583 (2019)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"17_CR22","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","volume":"82","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Ma, J.: Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 82, 28\u201342 (2022)","journal-title":"Inf. Fusion"},{"key":"17_CR23","doi-asserted-by":"publisher","first-page":"5413","DOI":"10.1109\/TMM.2022.3192661","volume":"25","author":"W Tang","year":"2023","unstructured":"Tang, W., He, F., Liu, Y.: YDTR: infrared and visible image fusion via Y-shape dynamic transformer. IEEE Trans. Multim. 25, 5413\u20135428 (2023)","journal-title":"IEEE Trans. Multim."},{"issue":"7","key":"17_CR24","doi-asserted-by":"publisher","first-page":"3159","DOI":"10.1109\/TCSVT.2023.3234340","volume":"33","author":"W Tang","year":"2023","unstructured":"Tang, W., He, F., Liu, Y., Duan, Y., Si, T.: DATFuse: infrared and visible image fusion via dual attention transformer. IEEE Trans. Circuits Syst. Video Technol. 33(7), 3159\u20133172 (2023)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"17_CR25","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 4\u20139 Dec 2017, Long Beach, CA, USA, pp. 5998\u20136008 (2017)"},{"key":"17_CR26","doi-asserted-by":"crossref","unstructured":"Vs, V., Valanarasu, J.M.J., Oza, P., Patel, V.M.: Image fusion transformer. In: 2022 IEEE International Conference on Image Processing, ICIP 2022, Bordeaux, France, 16\u201319 Oct 2022, pp. 3566\u20133570 (2022)","DOI":"10.1109\/ICIP46576.2022.9897280"},{"issue":"6","key":"17_CR27","doi-asserted-by":"publisher","first-page":"3360","DOI":"10.1109\/TCSVT.2021.3109895","volume":"32","author":"Z Wang","year":"2022","unstructured":"Wang, Z., Wang, J., Wu, Y., Xu, J., Zhang, X.: UNFusion: a unified multi-scale densely connected network for infrared and visible image fusion. IEEE Trans. Circuits Syst. Video Technol. 32(6), 3360\u20133374 (2022)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"4","key":"17_CR28","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"issue":"1","key":"17_CR29","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2022","unstructured":"Xu, H., Ma, J., Jiang, J., Guo, X., Ling, H.: U2Fusion: a unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 502\u2013518 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"17_CR30","unstructured":"Yang, J., et al.: Focal Self-attention for Local-Global Interactions in Vision Transformers (2021). CoRR arXiv:abs\/2107.00641"},{"issue":"8","key":"17_CR31","doi-asserted-by":"publisher","first-page":"10535","DOI":"10.1109\/TPAMI.2023.3261282","volume":"45","author":"X Zhang","year":"2023","unstructured":"Zhang, X., Demiris, Y.: Visible and infrared image fusion using deep learning. IEEE Trans. Pattern Anal. Mach. Intell. 45(8), 10535\u201310554 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"17_CR32","doi-asserted-by":"crossref","unstructured":"Zhang, X., Ye, P., Qiao, D., Zhao, J., Peng, S., Xiao, G.: Object fusion tracking based on visible and infrared images using fully convolutional Siamese networks. In: 22th International Conference on Information Fusion, FUSION 2019, Ottawa, ON, Canada, 2\u20135 July 2019, pp.\u00a01\u20138 (2019)","DOI":"10.23919\/FUSION43075.2019.9011253"},{"key":"17_CR33","doi-asserted-by":"crossref","unstructured":"Zhao, Z., et al.: CDDFuse: correlation-driven dual-branch feature decomposition for multi-modality image fusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5906\u20135916 (2023)","DOI":"10.1109\/CVPR52729.2023.00572"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-8685-5_17","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,2]],"date-time":"2024-11-02T05:19:35Z","timestamp":1730524775000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-8685-5_17"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,3]]},"ISBN":["9789819786848","9789819786855"],"references-count":33,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-8685-5_17","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,3]]},"assertion":[{"value":"3 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Urumqi","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2024.prcv.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}