{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T02:37:39Z","timestamp":1743043059734,"version":"3.40.3"},"publisher-location":"Cham","reference-count":70,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031770029"},{"type":"electronic","value":"9783031770036"}],"license":[{"start":{"date-parts":[[2024,11,17]],"date-time":"2024-11-17T00:00:00Z","timestamp":1731801600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,17]],"date-time":"2024-11-17T00:00:00Z","timestamp":1731801600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-77003-6_9","type":"book-chapter","created":{"date-parts":[[2024,11,16]],"date-time":"2024-11-16T04:54:14Z","timestamp":1731732854000},"page":"108-131","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Cross-Scale Bilevel Aggregation for\u00a0Multi-exposure Fusion via\u00a0Conditional Generative Adversarial Network"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-0994-7773","authenticated-orcid":false,"given":"Longchun","family":"Wang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9911-3950","authenticated-orcid":false,"given":"Mali","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Hai","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Taojun","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Qingming","family":"Leng","sequence":"additional","affiliation":[]},{"given":"Xiwei","family":"Dong","sequence":"additional","affiliation":[]},{"given":"Jingjuan","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Guangxing","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,17]]},"reference":[{"key":"9_CR1","doi-asserted-by":"publisher","first-page":"7","DOI":"10.1023\/B:VISI.0000016144.56397.1a","volume":"58","author":"M Aggarwal","year":"2004","unstructured":"Aggarwal, M., Ahuja, N.: Split aperture imaging for high dynamic range. Int. J. Comput. Vision 58, 7\u201317 (2004)","journal-title":"Int. J. Comput. Vision"},{"issue":"8","key":"9_CR2","doi-asserted-by":"publisher","first-page":"4686","DOI":"10.3390\/app13084686","volume":"13","author":"Y Ai","year":"2023","unstructured":"Ai, Y., et al.: Multi-scale feature fusion with attention mechanism based on CGAN network for infrared image colorization. Appl. Sci. 13(8), 4686 (2023)","journal-title":"Appl. Sci."},{"key":"9_CR3","unstructured":"Arjovsky, M., Chintala, S., Bottou, L.: Wasserstein generative adversarial networks. In: International Conference on Machine Learning, pp. 214\u2013223. PMLR (2017)"},{"issue":"4","key":"9_CR4","doi-asserted-by":"publisher","first-page":"2049","DOI":"10.1109\/TIP.2018.2794218","volume":"27","author":"J Cai","year":"2018","unstructured":"Cai, J., Gu, S., Zhang, L.: Learning a deep single image contrast enhancer from multi-exposure images. IEEE Trans. Image Process. 27(4), 2049\u20132062 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"9_CR5","doi-asserted-by":"publisher","first-page":"3098","DOI":"10.1109\/TIP.2021.3058764","volume":"30","author":"X Deng","year":"2021","unstructured":"Deng, X., Zhang, Y., Xu, M., Gu, S., Duan, Y.: Deep coupled feedback network for joint exposure fusion and image super-resolution. IEEE Trans. Image Process. 30, 3098\u20133112 (2021)","journal-title":"IEEE Trans. Image Process."},{"issue":"4","key":"9_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3072959.3073592","volume":"36","author":"M Gharbi","year":"2017","unstructured":"Gharbi, M., Chen, J., Barron, J.T., Hasinoff, S.W., Durand, F.: Deep bilateral learning for real-time image enhancement. ACM Trans. Graph. (TOG) 36(4), 1\u201312 (2017)","journal-title":"ACM Trans. Graph. (TOG)"},{"issue":"11","key":"9_CR7","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1145\/3422622","volume":"63","author":"I Goodfellow","year":"2020","unstructured":"Goodfellow, I., et al.: Generative adversarial networks. Commun. ACM 63(11), 139\u2013144 (2020)","journal-title":"Commun. ACM"},{"key":"9_CR8","doi-asserted-by":"crossref","unstructured":"Guo, C., et al.: Zero-reference deep curve estimation for low-light image enhancement. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1780\u20131789 (2020)","DOI":"10.1109\/CVPR42600.2020.00185"},{"issue":"5","key":"9_CR9","doi-asserted-by":"publisher","first-page":"744","DOI":"10.1016\/j.compeleceng.2011.07.012","volume":"37","author":"MBA Haghighat","year":"2011","unstructured":"Haghighat, M.B.A., Aghagolzadeh, A., Seyedarabi, H.: A non-reference image fusion metric based on mutual information of image features. Comput. Electr. Eng. 37(5), 744\u2013756 (2011)","journal-title":"Comput. Electr. Eng."},{"key":"9_CR10","doi-asserted-by":"publisher","first-page":"248","DOI":"10.1016\/j.inffus.2021.10.006","volume":"79","author":"D Han","year":"2022","unstructured":"Han, D., Li, L., Guo, X., Ma, J.: Multi-exposure image fusion via deep perceptual enhancement. Inf. Fusion 79, 248\u2013262 (2022)","journal-title":"Inf. Fusion"},{"key":"9_CR11","doi-asserted-by":"crossref","unstructured":"Han, K., Wang, Y., Tian, Q., Guo, J., Xu, C., Xu, C.: GhostNet: more features from cheap operations. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1580\u20131589 (2020)","DOI":"10.1109\/CVPR42600.2020.00165"},{"issue":"2","key":"9_CR12","doi-asserted-by":"publisher","first-page":"127","DOI":"10.1016\/j.inffus.2011.08.002","volume":"14","author":"Y Han","year":"2013","unstructured":"Han, Y., Cai, Y., Cao, Y., Xu, X.: A new image fusion performance metric based on visual information fidelity. Inf. fusion 14(2), 127\u2013135 (2013)","journal-title":"Inf. fusion"},{"issue":"14","key":"9_CR13","first-page":"7","volume":"43","author":"M Hassan","year":"2012","unstructured":"Hassan, M., Bhagvati, C.: Structural similarity measure for color images. Int. J. Comput. Appl. 43(14), 7\u201312 (2012)","journal-title":"Int. J. Comput. Appl."},{"issue":"7","key":"9_CR14","doi-asserted-by":"publisher","first-page":"1620","DOI":"10.3390\/math11071620","volume":"11","author":"CG Im","year":"2023","unstructured":"Im, C.G., Son, D.M., Kwon, H.J., Lee, S.H.: Multi-task learning approach using dynamic hyperparameter for multi-exposure fusion. Mathematics 11(7), 1620 (2023)","journal-title":"Mathematics"},{"key":"9_CR15","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"9_CR16","doi-asserted-by":"publisher","first-page":"305","DOI":"10.1016\/j.inffus.2022.07.013","volume":"88","author":"Z Le","year":"2022","unstructured":"Le, Z., et al.: UIFGAN: an unsupervised continual-learning generative adversarial network for unified image fusion. Inf. Fusion 88, 305\u2013318 (2022)","journal-title":"Inf. Fusion"},{"key":"9_CR17","doi-asserted-by":"crossref","unstructured":"Lee, S.H., Park, J.S., Cho, N.I.: A multi-exposure image fusion based on the adaptive weights reflecting the relative pixel intensity and global gradient. In: 2018 25th IEEE International Conference on Image Processing (ICIP), pp. 1737\u20131741. IEEE (2018)","DOI":"10.1109\/ICIP.2018.8451153"},{"key":"9_CR18","first-page":"1","volume":"72","author":"J Lei","year":"2023","unstructured":"Lei, J., Li, J., Liu, J., Zhou, S., Zhang, Q., Kasabov, N.K.: GALFusion: multi-exposure image fusion via a global-local aggregation learning network. IEEE Trans. Instrum. Meas. 72, 1\u201315 (2023)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"9_CR19","doi-asserted-by":"publisher","first-page":"5805","DOI":"10.1109\/TIP.2020.2987133","volume":"29","author":"H Li","year":"2020","unstructured":"Li, H., Ma, K., Yong, H., Zhang, L.: Fast multi-scale structural patch decomposition for multi-exposure image fusion. IEEE Trans. Image Process. 29, 5805\u20135816 (2020)","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"9_CR20","doi-asserted-by":"publisher","first-page":"713","DOI":"10.1109\/TCSVT.2022.3202692","volume":"33","author":"J Li","year":"2022","unstructured":"Li, J., Liu, J., Zhou, S., Zhang, Q., Kasabov, N.K.: Learning a coordinated network for detail-refinement multiexposure image fusion. IEEE Trans. Circuits Syst. Video Technol. 33(2), 713\u2013727 (2022)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"9_CR21","doi-asserted-by":"crossref","unstructured":"Li, L., Liang, D., Gao, Y., Huang, S.J., Chen, S.: ALL-E: aesthetics-guided low-light image enhancement. arXiv preprint arXiv:2304.14610 (2023)","DOI":"10.24963\/ijcai.2023\/118"},{"issue":"6","key":"9_CR22","doi-asserted-by":"publisher","first-page":"2828","DOI":"10.1109\/TIP.2018.2810539","volume":"27","author":"M Li","year":"2018","unstructured":"Li, M., Liu, J., Yang, W., Sun, X., Guo, Z.: Structure-revealing low-light image enhancement via robust retinex model. IEEE Trans. Image Process. 27(6), 2828\u20132841 (2018)","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"9_CR23","doi-asserted-by":"publisher","first-page":"626","DOI":"10.1109\/TCE.2012.6227469","volume":"58","author":"S Li","year":"2012","unstructured":"Li, S., Kang, X.: Fast multi-exposure image fusion with median filter and recursive filter. IEEE Trans. Consum. Electron. 58(2), 626\u2013632 (2012)","journal-title":"IEEE Trans. Consum. Electron."},{"key":"9_CR24","doi-asserted-by":"crossref","unstructured":"Li, S., Kang, X., Fang, L., Hu, J., Yin, H.: Pixel-level image fusion: a survey of the state of the art. Inf. Fusion 33, 100\u2013112 (2017)","DOI":"10.1016\/j.inffus.2016.05.004"},{"issue":"7","key":"9_CR25","doi-asserted-by":"publisher","first-page":"2864","DOI":"10.1109\/TIP.2013.2244222","volume":"22","author":"S Li","year":"2013","unstructured":"Li, S., Kang, X., Hu, J.: Image fusion with guided filtering. IEEE Trans. Image Process. 22(7), 2864\u20132875 (2013)","journal-title":"IEEE Trans. Image Process."},{"issue":"3","key":"9_CR26","doi-asserted-by":"publisher","first-page":"1243","DOI":"10.1109\/TIP.2017.2651366","volume":"26","author":"Z Li","year":"2017","unstructured":"Li, Z., Wei, Z., Wen, C., Zheng, J.: Detail-enhanced multi-scale exposure fusion. IEEE Trans. Image Process. 26(3), 1243\u20131252 (2017)","journal-title":"IEEE Trans. Image Process."},{"key":"9_CR27","doi-asserted-by":"publisher","first-page":"171139","DOI":"10.1016\/j.ijleo.2023.171139","volume":"288","author":"G Liu","year":"2023","unstructured":"Liu, G., Liu, Y., Tang, L., Bavirisetti, D.P., Wang, X.: A generative adversarial network for infrared and visible image fusion using adaptive dense generator and Markovian discriminator. Optik 288, 171139 (2023)","journal-title":"Optik"},{"issue":"8","key":"9_CR28","doi-asserted-by":"publisher","first-page":"5026","DOI":"10.1109\/TCSVT.2022.3144455","volume":"32","author":"J Liu","year":"2022","unstructured":"Liu, J., Shang, J., Liu, R., Fan, X.: Attention-guided global-local adversarial learning for detail-preserving multi-exposure image fusion. IEEE Trans. Circuits Syst. Video Technol. 32(8), 5026\u20135040 (2022)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"9_CR29","doi-asserted-by":"publisher","first-page":"237","DOI":"10.1016\/j.inffus.2023.02.027","volume":"95","author":"J Liu","year":"2023","unstructured":"Liu, J., Wu, G., Luan, J., Jiang, Z., Liu, R., Fan, X.: HoLoCo: holistic and local contrastive learning network for multi-exposure image fusion. Inf. Fusion 95, 237\u2013249 (2023)","journal-title":"Inf. Fusion"},{"key":"9_CR30","doi-asserted-by":"publisher","first-page":"208","DOI":"10.1016\/j.jvcir.2015.06.021","volume":"31","author":"Y Liu","year":"2015","unstructured":"Liu, Y., Wang, Z.: Dense sift for ghost-free multi-exposure fusion. J. Vis. Commun. Image Represent. 31, 208\u2013224 (2015)","journal-title":"J. Vis. Commun. Image Represent."},{"issue":"1","key":"9_CR31","doi-asserted-by":"publisher","first-page":"60","DOI":"10.1109\/TCI.2017.2786138","volume":"4","author":"K Ma","year":"2017","unstructured":"Ma, K., Duanmu, Z., Yeganeh, H., Wang, Z.: Multi-exposure image fusion by optimizing a structural similarity index. IEEE Trans. Comput. Imaging 4(1), 60\u201372 (2017)","journal-title":"IEEE Trans. Comput. Imaging"},{"key":"9_CR32","doi-asserted-by":"publisher","first-page":"2808","DOI":"10.1109\/TIP.2019.2952716","volume":"29","author":"K Ma","year":"2019","unstructured":"Ma, K., Duanmu, Z., Zhu, H., Fang, Y., Wang, Z.: Deep guided learning for fast multi-exposure image fusion. IEEE Trans. Image Process. 29, 2808\u20132819 (2019)","journal-title":"IEEE Trans. Image Process."},{"key":"9_CR33","doi-asserted-by":"crossref","unstructured":"Ma, K., Wang, Z.: Multi-exposure image fusion: a patch-wise approach. In: 2015 IEEE International Conference on Image Processing (ICIP), pp. 1717\u20131721. IEEE (2015)","DOI":"10.1109\/ICIP.2015.7351094"},{"issue":"11","key":"9_CR34","doi-asserted-by":"publisher","first-page":"3345","DOI":"10.1109\/TIP.2015.2442920","volume":"24","author":"K Ma","year":"2015","unstructured":"Ma, K., Zeng, K., Wang, Z.: Perceptual quality assessment for multi-exposure image fusion. IEEE Trans. Image Process. 24(11), 3345\u20133356 (2015)","journal-title":"IEEE Trans. Image Process."},{"issue":"5","key":"9_CR35","doi-asserted-by":"publisher","first-page":"582","DOI":"10.3390\/e24050582","volume":"24","author":"X Ma","year":"2022","unstructured":"Ma, X., Wang, Z., Hu, S., Kan, S.: Multi-focus image fusion based on multi-scale generative adversarial network. Entropy 24(5), 582 (2022)","journal-title":"Entropy"},{"key":"9_CR36","doi-asserted-by":"crossref","unstructured":"Mao, X., Li, Q., Xie, H., Lau, R.Y., Wang, Z., Paul\u00a0Smolley, S.: Least squares generative adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2794\u20132802 (2017)","DOI":"10.1109\/ICCV.2017.304"},{"key":"9_CR37","doi-asserted-by":"crossref","unstructured":"Mertens, T., Kautz, J., Van\u00a0Reeth, F.: Exposure fusion. In: 15th Pacific Conference on Computer Graphics and Applications (PG 2007), pp. 382\u2013390. IEEE (2007)","DOI":"10.1109\/PG.2007.17"},{"key":"9_CR38","unstructured":"Mirza, M., Osindero, S.: Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784 (2014)"},{"key":"9_CR39","doi-asserted-by":"crossref","unstructured":"Piella, G., Heijmans, H.: A new quality metric for image fusion. In: Proceedings 2003 International Conference on Image Processing (Cat. No. 03CH37429), vol.\u00a03, pp. III\u2013173. IEEE (2003)","DOI":"10.1109\/ICIP.2003.1247209"},{"key":"9_CR40","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1016\/j.inffus.2020.08.012","volume":"66","author":"Y Qi","year":"2021","unstructured":"Qi, Y., et al.: Deep unsupervised learning based on color un-referenced loss functions for multi-exposure image fusion. Inf. Fusion 66, 18\u201339 (2021)","journal-title":"Inf. Fusion"},{"key":"9_CR41","doi-asserted-by":"crossref","unstructured":"Qu, L., Liu, S., Wang, M., Song, Z.: TransMEF: a transformer-based multi-exposure image fusion framework using self-supervised multi-task learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a036, pp. 2126\u20132134 (2022)","DOI":"10.1609\/aaai.v36i2.20109"},{"key":"9_CR42","unstructured":"Radford, A., Metz, L., Chintala, S.: Unsupervised representation learning with deep convolutional generative adversarial networks. arxiv 2015. arXiv preprint arXiv:1511.06434 (2015)"},{"key":"9_CR43","doi-asserted-by":"crossref","unstructured":"Ram\u00a0Prabhakar, K., Sai\u00a0Srikar, V., Venkatesh\u00a0Babu, R.: DeepFuse: a deep unsupervised approach for exposure fusion with extreme exposure image pairs. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4714\u20134722 (2017)","DOI":"10.1109\/ICCV.2017.505"},{"key":"9_CR44","doi-asserted-by":"crossref","unstructured":"Raman, S., Chaudhuri, S.: Bilateral filter based compositing for variable exposure photography. In: Eurographics (Short Papers), pp.\u00a01\u20134 (2009)","DOI":"10.1145\/1666778.1666822"},{"key":"9_CR45","doi-asserted-by":"publisher","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W., Frangi, A. (eds.) Medical Image Computing and Computer-Assisted Intervention\u2013MICCAI 2015: 18th International Conference, Munich, Germany, 5\u20139 October 2015, Proceedings, Part III 18, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28","DOI":"10.1007\/978-3-319-24574-4_28"},{"issue":"9","key":"9_CR46","doi-asserted-by":"publisher","first-page":"1579","DOI":"10.1109\/TCYB.2013.2290435","volume":"44","author":"J Shen","year":"2014","unstructured":"Shen, J., Zhao, Y., Yan, S., Li, X., et al.: Exposure fusion using boosting Laplacian pyramid. IEEE Trans. Cybern. 44(9), 1579\u20131590 (2014)","journal-title":"IEEE Trans. Cybern."},{"key":"9_CR47","first-page":"1","volume":"60","author":"Z Tan","year":"2021","unstructured":"Tan, Z., Gao, M., Li, X., Jiang, L.: A flexible reference-insensitive spatiotemporal fusion model for remote sensing images using conditional generative adversarial network. IEEE Trans. Geosci. Remote Sens. 60, 1\u201313 (2021)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"9_CR48","first-page":"9969","volume":"35","author":"Y Tang","year":"2022","unstructured":"Tang, Y., Han, K., Guo, J., Xu, C., Xu, C., Wang, Y.: GhostNetV2: enhance cheap operation with long-range attention. Adv. Neural. Inf. Process. Syst. 35, 9969\u20139982 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"9_CR49","doi-asserted-by":"crossref","unstructured":"Tumblin, J., Agrawal, A., Raskar, R.: Why i want a gradient camera. In: 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2005), vol.\u00a01, pp. 103\u2013110. IEEE (2005)","DOI":"10.1109\/CVPR.2005.374"},{"key":"9_CR50","unstructured":"Umer, R.M., Foresti, G.L., Micheloni, C.: Deep generative adversarial residual convolutional networks for real-world super-resolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 438\u2013439 (2020)"},{"issue":"5","key":"9_CR51","doi-asserted-by":"publisher","first-page":"1233","DOI":"10.1007\/s00371-021-02079-5","volume":"37","author":"C Wang","year":"2021","unstructured":"Wang, C., He, C., Xu, M.: Fast exposure fusion of detail enhancement for brightest and darkest regions. Vis. Comput. 37(5), 1233\u20131243 (2021). https:\/\/doi.org\/10.1007\/s00371-021-02079-5","journal-title":"Vis. Comput."},{"issue":"2","key":"9_CR52","doi-asserted-by":"publisher","first-page":"560","DOI":"10.1587\/transinf.2017EDL8173","volume":"101","author":"J Wang","year":"2018","unstructured":"Wang, J., Wang, W., Xu, G., Liu, H.: End-to-end exposure fusion using convolutional neural network. IEICE Trans. Inf. Syst. 101(2), 560\u2013563 (2018)","journal-title":"IEICE Trans. Inf. Syst."},{"issue":"4","key":"9_CR53","doi-asserted-by":"publisher","first-page":"2709","DOI":"10.3390\/app13042709","volume":"13","author":"JG Wang","year":"2023","unstructured":"Wang, J.G., Wu, C.S.: Multi-scale aggregation residual channel attention fusion network for single image deraining. Appl. Sci. 13(4), 2709 (2023)","journal-title":"Appl. Sci."},{"issue":"12","key":"9_CR54","doi-asserted-by":"publisher","first-page":"8874","DOI":"10.1109\/TPAMI.2021.3123686","volume":"44","author":"L Wang","year":"2021","unstructured":"Wang, L., Yoon, K.J.: Deep learning for HDR imaging: state-of-the-art and future trends. IEEE Trans. Pattern Anal. Mach. Intell. 44(12), 8874\u20138895 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"8","key":"9_CR55","doi-asserted-by":"publisher","first-page":"2418","DOI":"10.1109\/TCSVT.2019.2919310","volume":"30","author":"Q Wang","year":"2019","unstructured":"Wang, Q., Chen, W., Wu, X., Li, Z.: Detail-enhanced multi-scale exposure fusion in YUV color space. IEEE Trans. Circuits Syst. Video Technol. 30(8), 2418\u20132429 (2019)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"9_CR56","doi-asserted-by":"crossref","unstructured":"Wu, W., Weng, J., Zhang, P., Wang, X., Yang, W., Jiang, J.: URetinex-Net: retinex-based deep unfolding network for low-light image enhancement. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5901\u20135910 (2022)","DOI":"10.1109\/CVPR52688.2022.00581"},{"key":"9_CR57","doi-asserted-by":"crossref","unstructured":"Wu, X., Huang, T.Z., Deng, L.J., Zhang, T.J.: Dynamic cross feature fusion for remote sensing pansharpening. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14687\u201314696 (2021)","DOI":"10.1109\/ICCV48922.2021.01442"},{"issue":"3","key":"9_CR58","doi-asserted-by":"publisher","first-page":"771","DOI":"10.3390\/rs14030771","volume":"14","author":"F Xu","year":"2022","unstructured":"Xu, F., Liu, J., Song, Y., Sun, H., Wang, X.: Multi-exposure image fusion techniques: a comprehensive review. Remote Sens. 14(3), 771 (2022)","journal-title":"Remote Sens."},{"issue":"1","key":"9_CR59","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2020","unstructured":"Xu, H., Ma, J., Jiang, J., Guo, X., Ling, H.: U2Fusion: a unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 502\u2013518 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"9_CR60","doi-asserted-by":"publisher","first-page":"7203","DOI":"10.1109\/TIP.2020.2999855","volume":"29","author":"H Xu","year":"2020","unstructured":"Xu, H., Ma, J., Zhang, X.P.: MEF-GAN: multi-exposure image fusion via generative adversarial networks. IEEE Trans. Image Process. 29, 7203\u20137216 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"9_CR61","doi-asserted-by":"publisher","first-page":"165494","DOI":"10.1016\/j.ijleo.2020.165494","volume":"223","author":"Y Xu","year":"2020","unstructured":"Xu, Y., Sun, B.: Color-compensated multi-scale exposure fusion based on physical features. Optik 223, 165494 (2020)","journal-title":"Optik"},{"issue":"4","key":"9_CR62","doi-asserted-by":"publisher","first-page":"308","DOI":"10.1049\/el:20000267","volume":"36","author":"CS Xydeas","year":"2000","unstructured":"Xydeas, C.S., Petrovic, V., et al.: Objective image fusion performance measure. Electron. Lett. 36(4), 308\u2013309 (2000)","journal-title":"Electron. Lett."},{"key":"9_CR63","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2022.3220285","volume":"71","author":"Y Yang","year":"2022","unstructured":"Yang, Y., Zhang, D., Wan, W., Huang, S.: Multi-scale exposure fusion based on multi-visual feature measurement and detail enhancement representation. IEEE Trans. Instrum. Meas. 71, 1\u201314 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"9_CR64","doi-asserted-by":"publisher","first-page":"6133","DOI":"10.1007\/s00521-020-05387-4","volume":"33","author":"Z Yang","year":"2021","unstructured":"Yang, Z., Chen, Y., Le, Z., Ma, Y.: GANFuse: a novel multi-exposure image fusion method based on generative adversarial networks. Neural Comput. Appl. 33, 6133\u20136145 (2021)","journal-title":"Neural Comput. Appl."},{"key":"9_CR65","doi-asserted-by":"publisher","first-page":"2841","DOI":"10.1109\/TMM.2021.3089324","volume":"24","author":"JL Yin","year":"2021","unstructured":"Yin, J.L., Chen, B.H., Peng, Y.T.: Two exposure fusion using prior-aware generative adversarial network. IEEE Trans. Multimedia 24, 2841\u20132851 (2021)","journal-title":"IEEE Trans. Multimedia"},{"key":"9_CR66","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1016\/j.inffus.2021.06.008","volume":"76","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Xu, H., Tian, X., Jiang, J., Ma, J.: Image fusion meets deep learning: a survey and perspective. Inf. Fusion 76, 323\u2013336 (2021)","journal-title":"Inf. Fusion"},{"key":"9_CR67","doi-asserted-by":"publisher","first-page":"101895","DOI":"10.1016\/j.inffus.2023.101895","volume":"99","author":"J Zhang","year":"2023","unstructured":"Zhang, J., Luo, Y., Huang, J., Liu, Y., Ma, J.: Multi-exposure image fusion via perception enhanced structural patch decomposition. Inf. Fusion 99, 101895 (2023)","journal-title":"Inf. Fusion"},{"key":"9_CR68","doi-asserted-by":"publisher","first-page":"111","DOI":"10.1016\/j.inffus.2021.02.005","volume":"74","author":"X Zhang","year":"2021","unstructured":"Zhang, X.: Benchmarking and comparing multi-exposure image fusion algorithms. Inf. Fusion 74, 111\u2013131 (2021)","journal-title":"Inf. Fusion"},{"key":"9_CR69","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","volume":"54","author":"Y Zhang","year":"2020","unstructured":"Zhang, Y., Liu, Y., Sun, P., Yan, H., Zhao, X., Zhang, L.: IFCNN: a general image fusion framework based on convolutional neural network. Inf. Fusion 54, 99\u2013118 (2020)","journal-title":"Inf. Fusion"},{"issue":"6","key":"9_CR70","first-page":"1433","volume":"3","author":"J Zhao","year":"2007","unstructured":"Zhao, J., Laganiere, R., Liu, Z.: Performance assessment of combinative pixel-level image fusion based on an absolute feature measurement. Int. J. Innov. Comput. Inf. Control 3(6), 1433\u20131447 (2007)","journal-title":"Int. J. Innov. Comput. Inf. Control"}],"container-title":["Lecture Notes in Computer Science","Internet of Things \u2013 ICIOT 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-77003-6_9","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,16]],"date-time":"2024-11-16T06:05:25Z","timestamp":1731737125000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-77003-6_9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,17]]},"ISBN":["9783031770029","9783031770036"],"references-count":70,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-77003-6_9","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,17]]},"assertion":[{"value":"17 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Disclosure of Interests"}},{"value":"SCF 2024 - ICIOT","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Internet of Things","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Bangkok","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Thailand","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16 November 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"19 November 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iciot2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.servicessociety.org\/iciot","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}