{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T16:38:45Z","timestamp":1769186325827,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":28,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819555666","type":"print"},{"value":"9789819555673","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-5567-3_18","type":"book-chapter","created":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:13:31Z","timestamp":1769116411000},"page":"256-270","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["HeRIF: A Mixture-of-Experts Framework for\u00a0Infrared and\u00a0Visible Image Fusion with\u00a0Heterogeneous Resolutions"],"prefix":"10.1007","author":[{"given":"Songqian","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Weijian","family":"Su","sequence":"additional","affiliation":[]},{"given":"Lei","family":"Meng","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8097-1397","authenticated-orcid":false,"given":"Yuqi","family":"Han","sequence":"additional","affiliation":[]},{"given":"Jinli","family":"Suo","sequence":"additional","affiliation":[]},{"given":"Qiang","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,23]]},"reference":[{"issue":"6","key":"18_CR1","doi-asserted-by":"publisher","first-page":"1391","DOI":"10.1109\/TGRS.2005.846874","volume":"43","author":"Z Wang","year":"2005","unstructured":"Wang, Z., Ziou, D., Armenakis, C., Li, D., Li, Q.: A comparative analysis of image fusion methods. IEEE Trans. Geosci. Remote Sens. 43(6), 1391\u20131402 (2005)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"18_CR2","doi-asserted-by":"crossref","unstructured":"Meng, Z., et al.: CoMoFusion: fast and high-quality fusion of infrared and visible image with consistency model. In: Chinese Conference on Pattern Recognition and Computer Vision, pp. 539\u2013553. Springer (2024)","DOI":"10.1007\/978-981-97-8685-5_38"},{"key":"18_CR3","doi-asserted-by":"crossref","unstructured":"Chen, M., Cheng, Y., He, X., Wang, X., Aze, Y., Xiang, J.: SimpleFusion: a simple fusion framework for infrared and visible images. In: Chinese Conference on Pattern Recognition and Computer Vision, pp. 49\u201363. Springer (2024)","DOI":"10.1007\/978-981-97-8685-5_4"},{"key":"18_CR4","doi-asserted-by":"crossref","unstructured":"Xu, H., Ma, J., Le, Z., Jiang, J., Guo, X.: FusionDN: a unified densely connected network for image fusion. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 12484\u201312491 (2020)","DOI":"10.1609\/aaai.v34i07.6936"},{"key":"18_CR5","doi-asserted-by":"publisher","first-page":"3845","DOI":"10.1109\/TIP.2020.2966075","volume":"29","author":"H Jung","year":"2020","unstructured":"Jung, H., Kim, Y., Jang, H., Ha, N., Sohn, K.: Unsupervised deep image fusion with structure tensor representations. IEEE Trans. Image Process. 29, 3845\u20133858 (2020)","journal-title":"IEEE Trans. Image Process."},{"issue":"1","key":"18_CR6","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"X Han","year":"2022","unstructured":"Han, X., Ma, J., Jiang, J., Guo, X., Ling, H.: U2Fusion: a unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 502\u2013518 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"18_CR7","doi-asserted-by":"crossref","unstructured":"Zhang, H., Han, X., Xiao, Y., Guo, X., Ma, J.: Rethinking the image fusion: a fast unified image fusion network based on proportional maintenance of gradient and intensity. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 12797\u201312804 (2020)","DOI":"10.1609\/aaai.v34i07.6975"},{"issue":"7","key":"18_CR8","doi-asserted-by":"publisher","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","volume":"9","author":"J Ma","year":"2022","unstructured":"Ma, J., Tang, L., Fan, F., Huang, J., Mei, X., Ma, Y.: SwinFusion: cross-domain long-range learning for general image fusion via swin transformer. IEEE\/CAA J. Automatica Sinica 9(7), 1200\u20131217 (2022)","journal-title":"IEEE\/CAA J. Automatica Sinica"},{"key":"18_CR9","doi-asserted-by":"crossref","unstructured":"Liang, P., Jiang, J., Liu, X., Ma, J.: Fusion from decomposition: a self-supervised decomposition approach for image fusion. In: European Conference on Computer Vision, pp. 719\u2013735. Springer (2022)","DOI":"10.1007\/978-3-031-19797-0_41"},{"key":"18_CR10","doi-asserted-by":"crossref","unstructured":"Zhang, H., Ma, J.: SDNet: a versatile squeeze-and-decomposition network for real-time image fusion. Int. J. Comput. Vis. 1\u201325 (2021)","DOI":"10.1007\/s11263-021-01501-8"},{"key":"18_CR11","doi-asserted-by":"publisher","first-page":"5413","DOI":"10.1109\/TMM.2022.3192661","volume":"25","author":"W Tang","year":"2023","unstructured":"Tang, W., He, F., Liu, Yu.: YDTR: Infrared and visible image fusion via y-shape dynamic transformer. IEEE Trans. Multimedia 25, 5413\u20135428 (2023)","journal-title":"IEEE Trans. Multimedia"},{"key":"18_CR12","unstructured":"Zhang, T., Zhao, J., Zhu, Y., Cui, G., Jing, Y., Lyu, Y.: DAAF: degradation-aware adaptive fusion framework for robust infrared and visible images fusion. arXiv:2504.10871 (2025)"},{"key":"18_CR13","doi-asserted-by":"crossref","unstructured":"Arnold, C., Jouvet, P., Seoud, L.: SwinFuSR: an image fusion-inspired model for RGB-guided thermal image super-resolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 3027\u20133036, June 2024","DOI":"10.1109\/CVPRW63382.2024.00308"},{"key":"18_CR14","doi-asserted-by":"crossref","unstructured":"Yi, X., Xu, H., Zhang, H., Tang, L., Ma, J.: Text-IF: leveraging semantic text guidance for degradation-aware and interactive image fusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 27026\u201327035 (2024)","DOI":"10.1109\/CVPR52733.2024.02552"},{"issue":"9","key":"18_CR15","doi-asserted-by":"publisher","first-page":"7577","DOI":"10.1109\/TPAMI.2025.3568433","volume":"47","author":"H Zhang","year":"2025","unstructured":"Zhang, H., Cao, L., Zuo, X., Shao, Z., Ma, J.: OmniFuse: composite degradation-robust image fusion with language-driven semantics. IEEE Trans. Pattern Anal. Mach. Intell. 47(9), 7577\u20137595 (2025)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"18_CR16","unstructured":"Luo, Y., et al.: WM-MoE: weather-aware multi-scale mixture-of-experts for blind adverse weather removal. arXiv preprint arXiv:2303.13739 (2023)"},{"key":"18_CR17","doi-asserted-by":"crossref","unstructured":"Cao, B., Sun, Y., Zhu, P., Hu, Q.: Multi-modal gated mixture of local-to-global experts for dynamic image fusion. In: Proceedings of IEEE\/CVF International Conference on Computer Vision, pp. 23498\u201323507 (2023)","DOI":"10.1109\/ICCV51070.2023.02153"},{"key":"18_CR18","doi-asserted-by":"crossref","unstructured":"Zhu, P., Sun, Y., Cao, B., Hu, Q.: Task-customized mixture of adapters for general image fusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7099\u20137108 (2024)","DOI":"10.1109\/CVPR52733.2024.00678"},{"key":"18_CR19","first-page":"13522","volume":"37","author":"H Guo","year":"2024","unstructured":"Guo, H., et al.: Parameter efficient adaptation for image restoration with heterogeneous mixture-of-experts. Adv. Neural. Inf. Process. Syst. 37, 13522\u201313547 (2024)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"18_CR20","doi-asserted-by":"crossref","unstructured":"Shin, W., Ahn, N., Moon, J.H., Sohn, K.A.: Exploiting distortion information for multi-degraded image restoration. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition workshop, pp. 536\u2013545 (2022)","DOI":"10.1109\/CVPRW56347.2022.00069"},{"key":"18_CR21","doi-asserted-by":"crossref","unstructured":"Ren, Y., et al.: MoE-DiffIR: task-customized diffusion priors for universal compressed image restoration. In: Proceedings of European Conference on Computer Vision, pp. 116\u2013134. Springer (2024)","DOI":"10.1007\/978-3-031-72673-6_7"},{"key":"18_CR22","doi-asserted-by":"crossref","unstructured":"Jia, X., Zhu, C., Li, M., Tang, W., Zhou, W.: LLVIP: a visible-infrared paired dataset for low-light vision. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3496\u20133504 (2021)","DOI":"10.1109\/ICCVW54120.2021.00389"},{"key":"18_CR23","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1016\/j.inffus.2022.03.007","volume":"83\u201384","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Zhang, H., Jiang, X., Ma, J.: PIAFusion: a progressive infrared and visible image fusion network based on illumination aware. Inf. Fusion 83\u201384, 79\u201392 (2022)","journal-title":"Inf. Fusion"},{"issue":"5","key":"18_CR24","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li, H., Xiao-Jun, W.: DenseFuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"18_CR25","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","volume":"73","author":"H Li","year":"2021","unstructured":"Li, H., Xiao-Jun, W., Kittler, J.: RFN-Nest: an end-to-end residual fusion network for infrared and visible images. Inf. Fusion 73, 72\u201386 (2021)","journal-title":"Inf. Fusion"},{"issue":"9","key":"18_CR26","doi-asserted-by":"publisher","first-page":"11040","DOI":"10.1109\/TPAMI.2023.3268209","volume":"45","author":"H Li","year":"2023","unstructured":"Li, H., Tianyang, X., Xiao-Jun, W., Jiwen, L., Kittler, J.: LRRNet: a novel representation learning guided fusion network for infrared and visible images. IEEE Trans. Pattern Anal. Mach. Intell. 45(9), 11040\u201311052 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"18_CR27","doi-asserted-by":"crossref","unstructured":"Wang, W., Deng, L.J., Vivone, G.: A general image fusion framework using multi-task semi-supervised learning. Inf. Fusion 102414 (2024)","DOI":"10.1016\/j.inffus.2024.102414"},{"key":"18_CR28","doi-asserted-by":"crossref","unstructured":"Yi, X., Tang, L., Zhang, H., Xu, H., Ma, J.: Diff-IF: multi-modality image fusion via diffusion model with fusion knowledge prior. Inf. Fusion 102450 (2024)","DOI":"10.1016\/j.inffus.2024.102450"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-5567-3_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:13:34Z","timestamp":1769116414000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-5567-3_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819555666","9789819555673"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-5567-3_18","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"23 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shanghai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2025.prcv.cn\/index.asp","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}