{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T14:51:28Z","timestamp":1743000688151,"version":"3.40.3"},"publisher-location":"Cham","reference-count":31,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031442223"},{"type":"electronic","value":"9783031442230"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-44223-0_39","type":"book-chapter","created":{"date-parts":[[2023,9,21]],"date-time":"2023-09-21T05:01:41Z","timestamp":1695272501000},"page":"485-496","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["MFT: Multi-scale Fusion Transformer for\u00a0Infrared and\u00a0Visible Image Fusion"],"prefix":"10.1007","author":[{"given":"Chen-Ming","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Chengbo","family":"Yuan","sequence":"additional","affiliation":[]},{"given":"Yong","family":"Luo","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Zhou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,9,22]]},"reference":[{"key":"39_CR1","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1016\/j.infrared.2017.11.006","volume":"88","author":"N Paramanandham","year":"2018","unstructured":"Paramanandham, N., Rajendiran, K.: Infrared and visible image fusion using discrete cosine transform and swarm intelligence for surveillance applications. Infrared Phys. Technol. 88, 13\u201322 (2018)","journal-title":"Infrared Phys. Technol."},{"issue":"9","key":"39_CR2","doi-asserted-by":"publisher","first-page":"4224","DOI":"10.1109\/TII.2018.2822828","volume":"14","author":"H Gao","year":"2018","unstructured":"Gao, H., Cheng, B., Wang, J., et al.: Object classification using CNN-based fusion of vision and LIDAR in autonomous vehicle environment. IEEE Trans. Industr. Inf. 14(9), 4224\u20134231 (2018)","journal-title":"IEEE Trans. Industr. Inf."},{"key":"39_CR3","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1016\/j.inffus.2018.02.004","volume":"45","author":"J Ma","year":"2019","unstructured":"Ma, J., Ma, Y., Li, C.: Infrared and visible image fusion methods and applications: a survey. Inf. Fusion 45, 153\u2013178 (2019)","journal-title":"Inf. Fusion"},{"key":"39_CR4","unstructured":"Kristan, M., Matas, J., Leonardis, A., et al.: The seventh visual object tracking VOT2019 challenge results. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision Workshops (2019)"},{"key":"39_CR5","doi-asserted-by":"publisher","first-page":"136","DOI":"10.1016\/j.inffus.2018.03.004","volume":"44","author":"C Lopez-Molina","year":"2018","unstructured":"Lopez-Molina, C., Montero, J., Bustince, H., et al.: Self-adapting weighted operators for multiscale gradient fusion. Inf. Fusion 44, 136\u2013146 (2018)","journal-title":"Inf. Fusion"},{"issue":"2","key":"39_CR6","doi-asserted-by":"publisher","first-page":"210","DOI":"10.1109\/TPAMI.2008.79","volume":"31","author":"J Wright","year":"2008","unstructured":"Wright, J., Yang, A.Y., Ganesh, A., et al.: Robust face recognition via sparse representation. IEEE Trans. Pattern Anal. Mach. Intell. 31(2), 210\u2013227 (2008)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"1","key":"39_CR7","doi-asserted-by":"publisher","first-page":"015011","DOI":"10.1117\/1.JRS.11.015011","volume":"11","author":"K He","year":"2017","unstructured":"He, K., Zhou, D., Zhang, X., et al.: Infrared and visible image fusion based on target extraction in the nonsubsampled contourlet transform domain. J. Appl. Remote Sens. 11(1), 015011\u2013015011 (2017)","journal-title":"J. Appl. Remote Sens."},{"issue":"1","key":"39_CR8","doi-asserted-by":"publisher","first-page":"171","DOI":"10.1109\/TPAMI.2012.88","volume":"35","author":"G Liu","year":"2012","unstructured":"Liu, G., Lin, Z., Yan, S., et al.: Robust recovery of subspace structures by low-rank representation. IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 171\u2013184 (2012)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"39_CR9","doi-asserted-by":"publisher","first-page":"94","DOI":"10.1016\/j.infrared.2017.04.018","volume":"83","author":"CH Liu","year":"2017","unstructured":"Liu, C.H., Qi, Y., Ding, W.R.: Infrared and visible image fusion method based on saliency detection in sparse domain. Infrared Phys. Technol. 83, 94\u2013102 (2017)","journal-title":"Infrared Phys. Technol."},{"issue":"5","key":"39_CR10","doi-asserted-by":"publisher","first-page":"057006","DOI":"10.1117\/1.OE.52.5.057006","volume":"52","author":"Q Zhang","year":"2013","unstructured":"Zhang, Q., Fu, Y., Li, H., et al.: Dictionary learning method for joint sparse representation-based image fusion. Opt. Eng. 52(5), 057006\u2013057006 (2013)","journal-title":"Opt. Eng."},{"key":"39_CR11","doi-asserted-by":"crossref","unstructured":"Li, H., Wu, X.J., Kittler, J.: Infrared and visible image fusion using a deep learning framework. In: 2018 24th International Conference on Pattern Recognition (ICPR), pp. 2705\u20132710. IEEE (2018)","DOI":"10.1109\/ICPR.2018.8546006"},{"key":"39_CR12","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"issue":"5","key":"39_CR13","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li, H., Wu, X.J.: DenseFuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"39_CR14","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., et al.: FusionGAN: a generative adversarial network for infrared and visible image fusion. Inf. Fusion 48, 11\u201326 (2019)","journal-title":"Inf. Fusion"},{"key":"39_CR15","doi-asserted-by":"crossref","unstructured":"Hu, R., Singh, A.: Transformer is all you need: Multimodal multitask learning with a unified transformer. arXiv preprint arXiv:2102.10772 (2021)","DOI":"10.1109\/ICCV48922.2021.00147"},{"key":"39_CR16","doi-asserted-by":"crossref","unstructured":"Chen, C.F.R., Fan, Q., Panda, R.: CrossViT: cross-attention multi-scale vision transformer for image classification. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 357\u2013366 (2021)","DOI":"10.1109\/ICCV48922.2021.00041"},{"key":"39_CR17","unstructured":"Ho, J., Kalchbrenner, N., Weissenborn, D., et al.: Axial attention in multidimensional transformers. arXiv preprint arXiv:1912.12180 (2019)"},{"key":"39_CR18","doi-asserted-by":"crossref","unstructured":"Fan, H., Xiong, B., Mangalam, K., et al.: Multiscale vision transformers. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6824\u20136835 (2021)","DOI":"10.1109\/ICCV48922.2021.00675"},{"key":"39_CR19","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","volume":"54","author":"Y Zhang","year":"2020","unstructured":"Zhang, Y., Liu, Y., Sun, P., et al.: IFCNN: a general image fusion framework based on convolutional neural network. Inf. Fusion 54, 99\u2013118 (2020)","journal-title":"Inf. Fusion"},{"issue":"12","key":"39_CR20","doi-asserted-by":"publisher","first-page":"9645","DOI":"10.1109\/TIM.2020.3005230","volume":"69","author":"H Li","year":"2020","unstructured":"Li, H., Wu, X.J., Durrani, T.: NestFuse: an infrared and visible image fusion architecture based on nest connection and spatial\/channel attention models. IEEE Trans. Instrum. Meas. 69(12), 9645\u20139656 (2020)","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"1","key":"39_CR21","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2020","unstructured":"Xu, H., Ma, J., Jiang, J., et al.: U2Fusion: a unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 502\u2013518 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"39_CR22","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","volume":"73","author":"H Li","year":"2021","unstructured":"Li, H., Wu, X.J., Kittler, J.: RFN-Nest: an end-to-end residual fusion network for infrared and visible images. Inf. Fusion 73, 72\u201386 (2021)","journal-title":"Inf. Fusion"},{"key":"39_CR23","doi-asserted-by":"crossref","unstructured":"Vs, V., Valanarasu, J.M.J., Oza, P., et al.: Image fusion transformer. In: 2022 IEEE International Conference on Image Processing (ICIP), pp. 3566\u20133570. IEEE (2022)","DOI":"10.1109\/ICIP46576.2022.9897280"},{"key":"39_CR24","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-10602-1_48","volume-title":"Computer Vision \u2013 ECCV 2014","author":"T-Y Lin","year":"2014","unstructured":"Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48"},{"key":"39_CR25","doi-asserted-by":"crossref","unstructured":"Hwang, S., Park, J., Kim, N., et al.: Multispectral pedestrian detection: benchmark dataset and baseline. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1037\u20131045 (2015)","DOI":"10.1109\/CVPR.2015.7298706"},{"key":"39_CR26","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1016\/j.dib.2017.09.038","volume":"15","author":"A Toet","year":"2017","unstructured":"Toet, A.: The TNO multiband image data collection. Data Brief 15, 249\u2013251 (2017)","journal-title":"Data Brief"},{"key":"39_CR27","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-030-00889-5_1","volume-title":"Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support","author":"Z Zhou","year":"2018","unstructured":"Zhou, Z., Rahman Siddiquee, M.M., Tajbakhsh, N., Liang, J.: UNet++: a nested U-Net architecture for medical image segmentation. In: Stoyanov, D., et al. (eds.) DLMIA\/ML-CDS -2018. LNCS, vol. 11045, pp. 3\u201311. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-00889-5_1"},{"key":"39_CR28","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"547","DOI":"10.1007\/978-3-030-68238-5_39","volume-title":"Computer Vision \u2013 ECCV 2020 Workshops","author":"M Kristan","year":"2020","unstructured":"Kristan, M., et al.: The eighth visual object tracking VOT2020 challenge results. In: Bartoli, A., Fusiello, A. (eds.) ECCV 2020. LNCS, vol. 12539, pp. 547\u2013601. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-68238-5_39"},{"key":"39_CR29","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"466","DOI":"10.1007\/978-3-031-19806-9_27","volume-title":"ECCV 2022","author":"Q Zhang","year":"2022","unstructured":"Zhang, Q., Xu, Y., Zhang, J., et al.: VSA: learning varied-size window attention in vision transformers. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13685, pp. 466\u2013483. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19806-9_27"},{"key":"39_CR30","first-page":"28522","volume":"34","author":"Y Xu","year":"2021","unstructured":"Xu, Y., Zhang, Q., Zhang, J., et al.: Vitae: vision transformer advanced by exploring intrinsic inductive bias. Adv. Neural. Inf. Process. Syst. 34, 28522\u201328535 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"39_CR31","doi-asserted-by":"crossref","unstructured":"Zhang, Q., Xu, Y., Zhang, J., et al.: ViTAEv2: vision transformer advanced by exploring inductive bias for image recognition and beyond. Int. J. Comput. Vision 1\u201322 (2023)","DOI":"10.1007\/s11263-022-01739-w"}],"container-title":["Lecture Notes in Computer Science","Artificial Neural Networks and Machine Learning \u2013 ICANN 2023"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-44223-0_39","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,9,21]],"date-time":"2023-09-21T06:16:59Z","timestamp":1695277019000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-44223-0_39"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031442223","9783031442230"],"references-count":31,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-44223-0_39","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"22 September 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICANN","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Artificial Neural Networks","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Heraklion","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Greece","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 September 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"32","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icann2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/e-nns.org\/icann2023\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"easyacademia.org","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"947","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"426","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"22","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"45% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.4","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"type of other papers accepted  : 9 Abstract","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}