{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,28]],"date-time":"2025-03-28T09:17:26Z","timestamp":1743153446504,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":41,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819786848"},{"type":"electronic","value":"9789819786855"}],"license":[{"start":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T00:00:00Z","timestamp":1730592000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T00:00:00Z","timestamp":1730592000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-97-8685-5_22","type":"book-chapter","created":{"date-parts":[[2024,11,2]],"date-time":"2024-11-02T05:05:25Z","timestamp":1730523925000},"page":"309-323","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["DBIF: Dual-Branch Feature Extraction Network for Infrared and Visible Image Fusion"],"prefix":"10.1007","author":[{"given":"Haozhe","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Rongpu","family":"Cui","sequence":"additional","affiliation":[]},{"given":"Zhuohang","family":"Zheng","sequence":"additional","affiliation":[]},{"given":"Shaobing","family":"Gao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,3]]},"reference":[{"key":"22_CR1","doi-asserted-by":"crossref","unstructured":"Cao, Y., Guan, D., Huang, W., Yang, J., Cao, Y., Qiao, Y.: Pedestrian detection with unsupervised multispectral feature learning using deep neural networks. Inf. Fusion 46, 206\u2013217 (2019)","DOI":"10.1016\/j.inffus.2018.06.005"},{"key":"22_CR2","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1016\/j.ins.2019.08.066","volume":"508","author":"J Chen","year":"2020","unstructured":"Chen, J., Li, X., Luo, L., Mei, X., Ma, J.: Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf. Sci. 508, 64\u201378 (2020)","journal-title":"Inf. Sci."},{"issue":"5","key":"22_CR3","doi-asserted-by":"publisher","first-page":"743","DOI":"10.1109\/JSEN.2007.894926","volume":"7","author":"N Cvejic","year":"2007","unstructured":"Cvejic, N., Bull, D., Canagarajah, N.: Region-based multimodal image fusion using ICA bases. IEEE Sens. J. 7(5), 743\u2013751 (2007)","journal-title":"IEEE Sens. J."},{"issue":"5","key":"22_CR4","doi-asserted-by":"publisher","first-page":"1808","DOI":"10.1109\/JSTARS.2015.2489838","volume":"9","author":"M Eslami","year":"2015","unstructured":"Eslami, M., Mohammadzadeh, A.: Developing a spectral-based strategy for urban object detection from airborne hyperspectral TIR and visible data. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 9(5), 1808\u20131816 (2015)","journal-title":"IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens."},{"key":"22_CR5","unstructured":"Fu, Y., Xu, T., Wu, X., Kittler, J.: Ppt fusion: Pyramid patch transformer for a case study in image fusion (2021). arXiv:2107.13967"},{"issue":"9","key":"22_CR6","doi-asserted-by":"publisher","first-page":"4387","DOI":"10.1109\/TIP.2019.2908783","volume":"28","author":"SB Gao","year":"2019","unstructured":"Gao, S.B., Ren, Y.Z., Zhang, M., Li, Y.J.: Combining bottom-up and top-down visual mechanisms for color constancy under varying illumination. IEEE Trans. Image Process. 28(9), 4387\u20134400 (2019)","journal-title":"IEEE Trans. Image Process."},{"key":"22_CR7","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2024.123391","volume":"249","author":"S Gao","year":"2024","unstructured":"Gao, S., He, L., Li, Y.: Biologically inspired image invariance guided illuminant estimation using shallow and deep models. Expert Syst. Appl. 249, 123391 (2024)","journal-title":"Expert Syst. Appl."},{"key":"22_CR8","doi-asserted-by":"crossref","unstructured":"Ha, Q., Watanabe, K., Karasawa, T., Ushiku, Y., Harada, T.: MFNet: towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In: 2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5108\u20135115 (2017)","DOI":"10.1109\/IROS.2017.8206396"},{"key":"22_CR9","doi-asserted-by":"crossref","unstructured":"Li, C., Zhu, C., Huang, Y., Tang, J., Wang, L.: Cross-modal ranking with soft consistency and noisy labels for robust RGB-T tracking. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 808\u2013823 (2018)","DOI":"10.1007\/978-3-030-01261-8_49"},{"issue":"5","key":"22_CR10","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li, H., Wu, X.J.: DenseFuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2018)","journal-title":"IEEE Trans. Image Process."},{"issue":"12","key":"22_CR11","doi-asserted-by":"publisher","first-page":"9645","DOI":"10.1109\/TIM.2020.3005230","volume":"69","author":"H Li","year":"2020","unstructured":"Li, H., Wu, X.J., Durrani, T.: NestFuse: an infrared and visible image fusion architecture based on nest connection and spatial\/channel attention models. IEEE Trans. Instrum. Meas. 69(12), 9645\u20139656 (2020)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"22_CR12","doi-asserted-by":"publisher","first-page":"4733","DOI":"10.1109\/TIP.2020.2975984","volume":"29","author":"H Li","year":"2020","unstructured":"Li, H., Wu, X.J., Kittler, J.: MDLatLRR: a novel decomposition method for infrared and visible image fusion. IEEE Trans. Image Process. 29, 4733\u20134746 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"22_CR13","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","volume":"73","author":"H Li","year":"2021","unstructured":"Li, H., Wu, X.J., Kittler, J.: RFN-nest: an end-to-end residual fusion network for infrared and visible images. Inf. Fusion 73, 72\u201386 (2021)","journal-title":"Inf. Fusion"},{"key":"22_CR14","doi-asserted-by":"crossref","unstructured":"Liu, J., Fan, X., Huang, Z., Wu, G., Liu, R., Zhong, W., Luo, Z.: Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5802\u20135811 (2022)","DOI":"10.1109\/CVPR52688.2022.00571"},{"issue":"12","key":"22_CR15","doi-asserted-by":"publisher","first-page":"1882","DOI":"10.1109\/LSP.2016.2618776","volume":"23","author":"Y Liu","year":"2016","unstructured":"Liu, Y., Chen, X., Ward, R.K., Wang, Z.J.: Image fusion with convolutional sparse representation. IEEE Signal Process. Lett. 23(12), 1882\u20131886 (2016)","journal-title":"IEEE Signal Process. Lett."},{"key":"22_CR16","unstructured":"Luo, Z., Tian, Y.: Infrared road object detection based on improved YOLOv8. IAENG Int. J. Comput. Sci. 51(3) (2024)"},{"key":"22_CR17","doi-asserted-by":"publisher","first-page":"100","DOI":"10.1016\/j.inffus.2016.02.001","volume":"31","author":"J Ma","year":"2016","unstructured":"Ma, J., Chen, C., Li, C., Huang, J.: Infrared and visible image fusion via gradient transfer and total variation minimization. Inf. Fusion 31, 100\u2013109 (2016)","journal-title":"Inf. Fusion"},{"key":"22_CR18","doi-asserted-by":"publisher","first-page":"85","DOI":"10.1016\/j.inffus.2019.07.005","volume":"54","author":"J Ma","year":"2020","unstructured":"Ma, J., Liang, P., Yu, W., Chen, C., Guo, X., Wu, J., Jiang, J.: Infrared and visible image fusion via detail preserving adversarial learning. Inf. Fusion 54, 85\u201398 (2020)","journal-title":"Inf. Fusion"},{"key":"22_CR19","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1016\/j.inffus.2018.02.004","volume":"45","author":"J Ma","year":"2019","unstructured":"Ma, J., Ma, Y., Li, C.: Infrared and visible image fusion methods and applications: a survey. Inf. Fusion 45, 153\u2013178 (2019)","journal-title":"Inf. Fusion"},{"key":"22_CR20","first-page":"1","volume":"70","author":"J Ma","year":"2021","unstructured":"Ma, J., Tang, L., Xu, M., Zhang, H., Xiao, G.: STDFusionNet: an infrared and visible image fusion network based on salient target detection. IEEE Trans. Instrum. Meas. 70, 1\u201313 (2021)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"22_CR21","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., Li, C., Jiang, J.: FusionGAN: a generative adversarial network for infrared and visible image fusion. Inf. Fusion 48, 11\u201326 (2019)","journal-title":"Inf. Fusion"},{"key":"22_CR22","first-page":"1","volume":"70","author":"J Ma","year":"2020","unstructured":"Ma, J., Zhang, H., Shao, Z., Liang, P., Xu, H.: GANMcC: a generative adversarial network with multiclassification constraints for infrared and visible image fusion. IEEE Trans. Instrum. Meas. 70, 1\u201314 (2020)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"22_CR23","doi-asserted-by":"publisher","first-page":"8","DOI":"10.1016\/j.infrared.2017.02.005","volume":"82","author":"J Ma","year":"2017","unstructured":"Ma, J., Zhou, Z., Wang, B., Zong, H.: Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys. Technol. 82, 8\u201317 (2017)","journal-title":"Infrared Phys. Technol."},{"issue":"11","key":"22_CR24","doi-asserted-by":"publisher","first-page":"4357","DOI":"10.1109\/TCSVT.2020.3047935","volume":"31","author":"MJ Tan","year":"2020","unstructured":"Tan, M.J., Gao, S.B., Xu, W.Z., Han, S.C.: Visible-infrared image fusion based on early visual information processing mechanisms. IEEE Trans. Circuits Syst. Video Technol. 31(11), 4357\u20134369 (2020)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"22_CR25","doi-asserted-by":"publisher","first-page":"1100697","DOI":"10.3389\/fnbot.2022.1100697","volume":"16","author":"M Tan","year":"2023","unstructured":"Tan, M., Yuan, X., Liang, B., Han, S.: DRFnet: dynamic receptive field network for object detection and image recognition. Front. Neurorobot. 16, 1100697 (2023)","journal-title":"Front. Neurorobot."},{"key":"22_CR26","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1016\/j.inffus.2022.03.007","volume":"83","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Zhang, H., Jiang, X., Ma, J.: PIAFusion: a progressive infrared and visible image fusion network based on illumination aware. Inf. Fusion 83, 79\u201392 (2022)","journal-title":"Inf. Fusion"},{"issue":"7","key":"22_CR27","doi-asserted-by":"publisher","first-page":"3159","DOI":"10.1109\/TCSVT.2023.3234340","volume":"33","author":"W Tang","year":"2023","unstructured":"Tang, W., He, F., Liu, Y., Duan, Y., Si, T.: DATFuse: infrared and visible image fusion via dual attention transformer. IEEE Trans. Circuits Syst. Video Technol. 33(7), 3159\u20133172 (2023)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"1","key":"22_CR28","doi-asserted-by":"publisher","first-page":"010901","DOI":"10.1117\/1.OE.51.1.010901","volume":"51","author":"A Toet","year":"2012","unstructured":"Toet, A., Hogervorst, M.A.: Progress in color night vision. Opt. Eng. 51(1), 010901\u2013010901 (2012)","journal-title":"Opt. Eng."},{"key":"22_CR29","doi-asserted-by":"crossref","unstructured":"Vs, V., Valanarasu, J.M.J., Oza, P., Patel, V.M.: Image fusion transformer. In: 2022 IEEE International Conference on Image Processing (ICIP), pp. 3566\u20133570 (2022)","DOI":"10.1109\/ICIP46576.2022.9897280"},{"key":"22_CR30","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2022.3216413","volume":"71","author":"Z Wang","year":"2022","unstructured":"Wang, Z., Chen, Y., Shao, W., Li, H., Zhang, L.: SwinFuse: a residual Swin transformer fusion network for infrared and visible images. IEEE Trans. Instrum. Meas. 71, 1\u201312 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"4","key":"22_CR31","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"key":"22_CR32","doi-asserted-by":"crossref","unstructured":"Xing, C., Wang, M., Wang, Z., Duan, C., Liu, Y.: Diagonalized low-rank learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. PP(99), 1\u201312 (2021)","DOI":"10.1109\/TGRS.2021.3085672"},{"key":"22_CR33","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2022.103407","volume":"218","author":"H Xu","year":"2022","unstructured":"Xu, H., Gong, M., Tian, X., Huang, J., Ma, J.: CUFD: an encoder-decoder network for visible and infrared image fusion based on common and unique feature decomposition. Comput. Vis. Image Underst. 218, 103407 (2022)","journal-title":"Comput. Vis. Image Underst."},{"key":"22_CR34","doi-asserted-by":"crossref","unstructured":"Xu, H., Ma, J., Le, Z., Jiang, J., Guo, X.: Fusiondn: a unified densely connected network for image fusion. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol.\u00a034, pp. 12484\u201312491 (2020)","DOI":"10.1609\/aaai.v34i07.6936"},{"key":"22_CR35","doi-asserted-by":"crossref","unstructured":"Xue, S., Gao, S., Tan, M., He, Z., He, L.: How does color constancy affect target recognition and instance segmentation? In: Proceedings of the 29th ACM International Conference on Multimedia, pp. 5537\u20135545 (2021)","DOI":"10.1145\/3474085.3475690"},{"key":"22_CR36","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1016\/j.inffus.2021.06.008","volume":"76","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Xu, H., Tian, X., Jiang, J., Ma, J.: Image fusion meets deep learning: a survey and perspective. Inf. Fusion 76, 323\u2013336 (2021)","journal-title":"Inf. Fusion"},{"key":"22_CR37","doi-asserted-by":"crossref","unstructured":"Zhang, H., Xu, H., Xiao, Y., Guo, X., Ma, J.: Rethinking the image fusion: a fast unified image fusion network based on proportional maintenance of gradient and intensity. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a034, pp. 12797\u201312804 (2020)","DOI":"10.1609\/aaai.v34i07.6975"},{"issue":"3","key":"22_CR38","doi-asserted-by":"publisher","first-page":"1219","DOI":"10.1109\/TIP.2016.2516953","volume":"25","author":"XS Zhang","year":"2016","unstructured":"Zhang, X.S., Gao, S.B., Li, R.X., Du, X.Y., Li, C.Y., Li, Y.J.: A retinal mechanism inspired color constancy model. IEEE Trans. Image Process. 25(3), 1219\u20131232 (2016)","journal-title":"IEEE Trans. Image Process."},{"key":"22_CR39","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","volume":"54","author":"Y Zhang","year":"2020","unstructured":"Zhang, Y., Liu, Y., Sun, P., Yan, H., Zhao, X., Zhang, L.: IFCNN: a general image fusion framework based on convolutional neural network. Inf. Fusion 54, 99\u2013118 (2020)","journal-title":"Inf. Fusion"},{"key":"22_CR40","doi-asserted-by":"publisher","first-page":"189","DOI":"10.1016\/j.inffus.2021.06.002","volume":"76","author":"F Zhao","year":"2021","unstructured":"Zhao, F., Zhao, W., Yao, L., Liu, Y.: Self-supervised feature adaption for infrared and visible image fusion. Inf. Fusion 76, 189\u2013203 (2021)","journal-title":"Inf. Fusion"},{"key":"22_CR41","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Xu, S., Zhang, C., Liu, J., Li, P., Zhang, J.: DIDFuse: deep image decomposition for infrared and visible image fusion (2020). arXiv:2003.09210","DOI":"10.24963\/ijcai.2020\/135"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-8685-5_22","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,2]],"date-time":"2024-11-02T05:21:09Z","timestamp":1730524869000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-8685-5_22"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,3]]},"ISBN":["9789819786848","9789819786855"],"references-count":41,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-8685-5_22","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,3]]},"assertion":[{"value":"3 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Urumqi","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2024.prcv.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}