{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,28]],"date-time":"2025-03-28T01:55:32Z","timestamp":1743126932542,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":33,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819985487"},{"type":"electronic","value":"9789819985494"}],"license":[{"start":{"date-parts":[[2023,12,25]],"date-time":"2023-12-25T00:00:00Z","timestamp":1703462400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,12,25]],"date-time":"2023-12-25T00:00:00Z","timestamp":1703462400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-99-8549-4_7","type":"book-chapter","created":{"date-parts":[[2023,12,24]],"date-time":"2023-12-24T18:01:32Z","timestamp":1703440892000},"page":"77-88","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Infrared and\u00a0Visible Image Fusion via\u00a0Test-Time Training"],"prefix":"10.1007","author":[{"given":"Guoqing","family":"Zheng","sequence":"first","affiliation":[]},{"given":"Zhenqi","family":"Fu","sequence":"additional","affiliation":[]},{"given":"Xiaopeng","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Xueye","family":"Chu","sequence":"additional","affiliation":[]},{"given":"Yue","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Xinghao","family":"Ding","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,25]]},"reference":[{"issue":"12","key":"7_CR1","doi-asserted-by":"publisher","first-page":"1890","DOI":"10.1016\/j.aeue.2015.09.004","volume":"69","author":"V Aslantas","year":"2015","unstructured":"Aslantas, V., Bendes, E.: A new image quality metric for image fusion: the sum of the correlations of differences. Aeu-Inter. J. Electr. Commun. 69(12), 1890\u20131896 (2015)","journal-title":"Aeu-Inter. J. Electr. Commun."},{"issue":"1","key":"7_CR2","doi-asserted-by":"publisher","first-page":"40","DOI":"10.3141\/1708-05","volume":"1708","author":"S Das","year":"2000","unstructured":"Das, S., Zhang, Y.: Color night vision for navigation and surveillance. Transp. Res. Rec. 1708(1), 40\u201346 (2000)","journal-title":"Transp. Res. Rec."},{"key":"7_CR3","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255. Ieee (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"7_CR4","first-page":"29374","volume":"35","author":"Y Gandelsman","year":"2022","unstructured":"Gandelsman, Y., Sun, Y., Chen, X., Efros, A.: Test-time training with masked autoencoders. Adv. Neural. Inf. Process. Syst. 35, 29374\u201329385 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"7_CR5","doi-asserted-by":"crossref","unstructured":"Gao, Y., Ma, S., Liu, J.: Dcdr-gan: a densely connected disentangled representation generative adversarial network for infrared and visible image fusion. IEEE Trans. Circ. Syst. Video Technol. (2022)","DOI":"10.1109\/TCSVT.2022.3206807"},{"issue":"5","key":"7_CR6","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li, H., Wu, X.J.: Densefuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2018)","journal-title":"IEEE Trans. Image Process."},{"issue":"12","key":"7_CR7","doi-asserted-by":"publisher","first-page":"9645","DOI":"10.1109\/TIM.2020.3005230","volume":"69","author":"H Li","year":"2020","unstructured":"Li, H., Wu, X.J., Durrani, T.: Nestfuse: an infrared and visible image fusion architecture based on nest connection and spatial\/channel attention models. IEEE Trans. Instrum. Meas. 69(12), 9645\u20139656 (2020)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"7_CR8","doi-asserted-by":"crossref","unstructured":"Li, H., Wu, X.J., Kittler, J.: Infrared and visible image fusion using a deep learning framework. In: 2018 24th International Conference On Pattern Recognition (ICPR), pp. 2705\u20132710. IEEE (2018)","DOI":"10.1109\/ICPR.2018.8546006"},{"key":"7_CR9","doi-asserted-by":"publisher","first-page":"4733","DOI":"10.1109\/TIP.2020.2975984","volume":"29","author":"H Li","year":"2020","unstructured":"Li, H., Wu, X.J., Kittler, J.: Mdlatlrr: a novel decomposition method for infrared and visible image fusion. IEEE Trans. Image Process. 29, 4733\u20134746 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"7_CR10","first-page":"1","volume":"71","author":"Q Li","year":"2022","unstructured":"Li, Q., et al.: A multilevel hybrid transmission network for infrared and visible image fusion. IEEE Trans. Instrum. Meas. 71, 1\u201314 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"7_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-10602-1_48","volume-title":"Computer Vision \u2013 ECCV 2014","author":"T-Y Lin","year":"2014","unstructured":"Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48"},{"key":"7_CR12","first-page":"1","volume":"71","author":"X Lin","year":"2022","unstructured":"Lin, X., Zhou, G., Tu, X., Huang, Y., Ding, X.: Two-level consistency metric for infrared and visible image fusion. IEEE Trans. Instrum. Meas. 71, 1\u201313 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"1","key":"7_CR13","doi-asserted-by":"publisher","first-page":"171","DOI":"10.1109\/TPAMI.2012.88","volume":"35","author":"G Liu","year":"2012","unstructured":"Liu, G., Lin, Z., Yan, S., Sun, J., Yu, Y., Ma, Y.: Robust recovery of subspace structures by low-rank representation. IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 171\u2013184 (2012)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"7_CR14","unstructured":"Liu, G., Lin, Z., Yu, Y.: Robust subspace segmentation by low-rank representation. In: Proceedings of the 27th International Conference on Machine Learning (ICML 2010), pp. 663\u2013670 (2010)"},{"key":"7_CR15","doi-asserted-by":"crossref","unstructured":"Liu, H., Wu, Z., Li, L., Salehkalaibar, S., Chen, J., Wang, K.: Towards multi-domain single image dehazing via test-time training. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5831\u20135840 (2022)","DOI":"10.1109\/CVPR52688.2022.00574"},{"key":"7_CR16","doi-asserted-by":"publisher","first-page":"100","DOI":"10.1016\/j.inffus.2016.02.001","volume":"31","author":"J Ma","year":"2016","unstructured":"Ma, J., Chen, C., Li, C., Huang, J.: Infrared and visible image fusion via gradient transfer and total variation minimization. Inform. Fusion 31, 100\u2013109 (2016)","journal-title":"Inform. Fusion"},{"key":"7_CR17","doi-asserted-by":"publisher","first-page":"4980","DOI":"10.1109\/TIP.2020.2977573","volume":"29","author":"J Ma","year":"2020","unstructured":"Ma, J., Xu, H., Jiang, J., Mei, X., Zhang, X.P.: Ddcgan: a dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans. Image Process. 29, 4980\u20134995 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"7_CR18","unstructured":"Van der Maaten, L., Hinton, G.: Visualizing data using t-sne. J. Mach. Learn. Res. 9(11) (2008)"},{"issue":"4","key":"7_CR19","doi-asserted-by":"publisher","first-page":"259","DOI":"10.1016\/S1566-2535(03)00046-0","volume":"4","author":"G Piella","year":"2003","unstructured":"Piella, G.: A general framework for multiresolution image fusion: from pixels to regions. Inform. Fusion 4(4), 259\u2013280 (2003)","journal-title":"Inform. Fusion"},{"issue":"1","key":"7_CR20","doi-asserted-by":"publisher","DOI":"10.1117\/1.2945910","volume":"2","author":"JW Roberts","year":"2008","unstructured":"Roberts, J.W., Van Aardt, J.A., Ahmed, F.B.: Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J. Appl. Remote Sens. 2(1), 023522 (2008)","journal-title":"J. Appl. Remote Sens."},{"issue":"2","key":"7_CR21","doi-asserted-by":"publisher","first-page":"430","DOI":"10.1109\/TIP.2005.859378","volume":"15","author":"HR Sheikh","year":"2006","unstructured":"Sheikh, H.R., Bovik, A.C.: Image information and visual quality. IEEE Trans. Image Process. 15(2), 430\u2013444 (2006)","journal-title":"IEEE Trans. Image Process."},{"key":"7_CR22","unstructured":"Sun, Y., Wang, X., Liu, Z., Miller, J., Efros, A., Hardt, M.: Test-time training with self-supervision for generalization under distribution shifts. In: International Conference on Machine Learning, pp. 9229\u20139248. PMLR (2020)"},{"key":"7_CR23","doi-asserted-by":"crossref","unstructured":"Tang, W., He, F., Liu, Y.: Ydtr: infrared and visible image fusion via y-shape dynamic transformer. IEEE Trans. Multimedia (2022)","DOI":"10.1109\/TMM.2022.3192661"},{"key":"7_CR24","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1016\/j.dib.2017.09.038","volume":"15","author":"A Toet","year":"2017","unstructured":"Toet, A.: The tno multiband image data collection. Data Brief 15, 249\u2013251 (2017)","journal-title":"Data Brief"},{"issue":"9","key":"7_CR25","doi-asserted-by":"publisher","first-page":"3367","DOI":"10.1109\/TIM.2018.2877285","volume":"68","author":"A Vishwakarma","year":"2018","unstructured":"Vishwakarma, A.: Image fusion using adjustable non-subsampled shearlet transform. IEEE Trans. Instrum. Meas. 68(9), 3367\u20133378 (2018)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"7_CR26","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2022.3216413","volume":"71","author":"Z Wang","year":"2022","unstructured":"Wang, Z., Wu, Y., Wang, J., Xu, J., Shao, W.: Res2fusion: infrared and visible image fusion based on dense res2net and double nonlocal attention models. IEEE Trans. Instrum. Meas. 71, 1\u201312 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"4","key":"7_CR27","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"key":"7_CR28","unstructured":"Wang, Z., Simoncelli, E.P., Bovik, A.C.: Multiscale structural similarity for image quality assessment. In: The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers 2003, vol. 2, pp. 1398\u20131402. IEEE (2003)"},{"issue":"2","key":"7_CR29","doi-asserted-by":"publisher","first-page":"210","DOI":"10.1109\/TPAMI.2008.79","volume":"31","author":"J Wright","year":"2008","unstructured":"Wright, J., Yang, A.Y., Ganesh, A., Sastry, S.S., Ma, Y.: Robust face recognition via sparse representation. IEEE Trans. Pattern Anal. Mach. Intell. 31(2), 210\u2013227 (2008)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"1","key":"7_CR30","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2020","unstructured":"Xu, H., Ma, J., Jiang, J., Guo, X., Ling, H.: U2fusion: a unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 502\u2013518 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"5","key":"7_CR31","doi-asserted-by":"publisher","first-page":"057006","DOI":"10.1117\/1.OE.52.5.057006","volume":"52","author":"Q Zhang","year":"2013","unstructured":"Zhang, Q., Fu, Y., Li, H., Zou, J.: Dictionary learning method for joint sparse representation-based image fusion. Opt. Eng. 52(5), 057006\u2013057006 (2013)","journal-title":"Opt. Eng."},{"key":"7_CR32","doi-asserted-by":"crossref","unstructured":"Zhang, X., Demiris, Y.: Visible and infrared image fusion using deep learning. IEEE Trans. Pattern Anal. Mach. Intell. (2023)","DOI":"10.1109\/TPAMI.2023.3261282"},{"key":"7_CR33","doi-asserted-by":"crossref","unstructured":"Zhang, X., Ye, P., Xiao, G.: Vifb: a visible and infrared image fusion benchmark. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 104\u2013105 (2020)","DOI":"10.1109\/CVPRW50498.2020.00060"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-8549-4_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,24]],"date-time":"2023-12-24T18:08:03Z","timestamp":1703441283000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-8549-4_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,25]]},"ISBN":["9789819985487","9789819985494"],"references-count":33,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-8549-4_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023,12,25]]},"assertion":[{"value":"25 December 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Xiamen","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/prcv2023.xmu.edu.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1420","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"532","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"37% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,78","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,69","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}