{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T18:16:46Z","timestamp":1743099406982,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":31,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819600250"},{"type":"electronic","value":"9789819600267"}],"license":[{"start":{"date-parts":[[2024,11,16]],"date-time":"2024-11-16T00:00:00Z","timestamp":1731715200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,16]],"date-time":"2024-11-16T00:00:00Z","timestamp":1731715200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-0026-7_10","type":"book-chapter","created":{"date-parts":[[2024,11,15]],"date-time":"2024-11-15T19:00:59Z","timestamp":1731697259000},"page":"120-135","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["EBcGAN: An Edge-Based Conditional Generative Adversarial Network for\u00a0Image Fusion"],"prefix":"10.1007","author":[{"given":"Mengshu","family":"Li","sequence":"first","affiliation":[]},{"given":"Zheyuan","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Yuai","family":"Hua","sequence":"additional","affiliation":[]},{"given":"Jinyong","family":"Cheng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,16]]},"reference":[{"key":"10_CR1","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2024.3390194","volume":"73","author":"J Chen","year":"2024","unstructured":"Chen, J., Yang, L., Liu, W., Tian, X., Ma, J.: Lenfusion: a joint low-light enhancement and fusion network for nighttime infrared and visible image fusion. IEEE Trans. Instrum. Meas. 73, 1\u201315 (2024). https:\/\/doi.org\/10.1109\/TIM.2024.3390194","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"5","key":"10_CR2","doi-asserted-by":"publisher","first-page":"743","DOI":"10.1109\/JSEN.2007.894926","volume":"7","author":"N Cvejic","year":"2007","unstructured":"Cvejic, N., Bull, D., Canagarajah, N.: Region-based multimodal image fusion using ICA bases. IEEE Sens. J. 7(5), 743\u2013751 (2007). https:\/\/doi.org\/10.1109\/JSEN.2007.894926","journal-title":"IEEE Sens. J."},{"issue":"5","key":"10_CR3","doi-asserted-by":"publisher","first-page":"1808","DOI":"10.1109\/JSTARS.2015.2489838","volume":"9","author":"M Eslami","year":"2016","unstructured":"Eslami, M., Mohammadzadeh, A.: Developing a spectral-based strategy for urban object detection from airborne hyperspectral TIR and visible data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 9(5), 1808\u20131816 (2016). https:\/\/doi.org\/10.1109\/JSTARS.2015.2489838","journal-title":"IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens."},{"key":"10_CR4","doi-asserted-by":"publisher","unstructured":"Krishna, B.V.S., Rajalakshmi, B., Dhammini, U., Monika, M., Nethra, C., Ashok, K.: Image de-hazing techniques for vision based applications - a survey. In: 2023 International Conference for Advancement in Technology (ICONAT). pp.\u00a01\u20135 (2023). https:\/\/doi.org\/10.1109\/ICONAT57137.2023.10080156","DOI":"10.1109\/ICONAT57137.2023.10080156"},{"key":"10_CR5","doi-asserted-by":"publisher","unstructured":"Li, B., Peng, X., Wang, Z., Xu, J., Feng, D.: AOD-Net: all-in-one dehazing network. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 4780\u20134788 (2017). https:\/\/doi.org\/10.1109\/ICCV.2017.511","DOI":"10.1109\/ICCV.2017.511"},{"issue":"5","key":"10_CR6","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2019","unstructured":"Li, H., Wu, X.J.: Densefuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2019). https:\/\/doi.org\/10.1109\/TIP.2018.2887342","journal-title":"IEEE Trans. Image Process."},{"issue":"12","key":"10_CR7","doi-asserted-by":"publisher","first-page":"9645","DOI":"10.1109\/TIM.2020.3005230","volume":"69","author":"H Li","year":"2020","unstructured":"Li, H., Wu, X.J., Durrani, T.: Nestfuse: an infrared and visible image fusion architecture based on nest connection and spatial\/channel attention models. IEEE Trans. Instrum. Meas. 69(12), 9645\u20139656 (2020). https:\/\/doi.org\/10.1109\/TIM.2020.3005230","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10_CR8","doi-asserted-by":"publisher","unstructured":"Li, H., Wu, X.J., Kittler, J.: RFN-Nest: an end-to-end residual fusion network for infrared and visible images. Inform. Fusion 73, 72\u201386 (2021). https:\/\/doi.org\/10.1016\/j.inffus.2021.02.023. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1566253521000440","DOI":"10.1016\/j.inffus.2021.02.023"},{"key":"10_CR9","doi-asserted-by":"publisher","first-page":"131","DOI":"10.1016\/j.neucom.2017.01.006","volume":"235","author":"X Liu","year":"2017","unstructured":"Liu, X., Mei, W., Du, H.: Structure tensor and nonsubsampled shearlet transform based algorithm for CT and MRI image fusion. Neurocomputing 235, 131\u2013139 (2017). https:\/\/doi.org\/10.1016\/j.neucom.2017.01.006. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0925231217300103","journal-title":"Neurocomputing"},{"key":"10_CR10","doi-asserted-by":"publisher","first-page":"9","DOI":"10.1016\/j.sigpro.2013.10.010","volume":"97","author":"Y Liu","year":"2014","unstructured":"Liu, Y., Jin, J., Wang, Q., Shen, Y., Dong, X.: Region level based multi-focus image fusion using quaternion wavelet and normalized cut. Signal Process. 97, 9\u201330 (2014). https:\/\/doi.org\/10.1016\/j.sigpro.2013.10.010. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0165168413004015","journal-title":"Signal Process."},{"key":"10_CR11","doi-asserted-by":"publisher","first-page":"147","DOI":"10.1016\/j.inffus.2014.09.004","volume":"24","author":"Y Liu","year":"2015","unstructured":"Liu, Y., Liu, S., Wang, Z.: A general framework for image fusion based on multi-scale transform and sparse representation. Inform. Fusion 24, 147\u2013164 (2015). https:\/\/doi.org\/10.1016\/j.inffus.2014.09.004. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1566253514001043","journal-title":"Inform. Fusion"},{"key":"10_CR12","doi-asserted-by":"publisher","first-page":"136","DOI":"10.1016\/j.inffus.2018.03.004","volume":"44","author":"C Lopez-Molina","year":"2018","unstructured":"Lopez-Molina, C., Montero, J., Bustince, H., De Baets, B.: Self-adapting weighted operators for multiscale gradient fusion. Inform. Fusion 44, 136\u2013146 (2018). https:\/\/doi.org\/10.1016\/j.inffus.2018.03.004. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1566253517304736","journal-title":"Inform. Fusion"},{"key":"10_CR13","doi-asserted-by":"publisher","first-page":"82134","DOI":"10.1109\/ACCESS.2024.3412157","volume":"12","author":"Y Luo","year":"2024","unstructured":"Luo, Y., Luo, Z.: Infrared and visible image fusion algorithm based on improved residual swin transformer and sobel operators. IEEE Access 12, 82134\u201382145 (2024). https:\/\/doi.org\/10.1109\/ACCESS.2024.3412157","journal-title":"IEEE Access"},{"key":"10_CR14","doi-asserted-by":"publisher","first-page":"85","DOI":"10.1016\/j.inffus.2019.07.005","volume":"54","author":"J Ma","year":"2020","unstructured":"Ma, J., Liang, P., Yu, W., Chen, C., Guo, X., Wu, J., Jiang, J.: Infrared and visible image fusion via detail preserving adversarial learning. Inform. Fusion 54, 85\u201398 (2020). https:\/\/doi.org\/10.1016\/j.inffus.2019.07.005. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1566253519300314","journal-title":"Inform. Fusion"},{"key":"10_CR15","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2021.3075747","volume":"70","author":"J Ma","year":"2021","unstructured":"Ma, J., Tang, L., Xu, M., Zhang, H., Xiao, G.: STDFusionNet: an infrared and visible image fusion network based on salient target detection. IEEE Trans. Instrum. Meas. 70, 1\u201313 (2021). https:\/\/doi.org\/10.1109\/TIM.2021.3075747","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10_CR16","doi-asserted-by":"publisher","first-page":"4980","DOI":"10.1109\/TIP.2020.2977573","volume":"29","author":"J Ma","year":"2020","unstructured":"Ma, J., Xu, H., Jiang, J., Mei, X., Zhang, X.P.: DDcGAN: a dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans. Image Process. 29, 4980\u20134995 (2020). https:\/\/doi.org\/10.1109\/TIP.2020.2977573","journal-title":"IEEE Trans. Image Process."},{"key":"10_CR17","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., Li, C., Jiang, J.: FusionGAN: a generative adversarial network for infrared and visible image fusion. Inform. Fusion 48, 11\u201326 (2019). https:\/\/doi.org\/10.1016\/j.inffus.2018.09.004. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1566253518301143","journal-title":"Inform. Fusion"},{"key":"10_CR18","doi-asserted-by":"publisher","first-page":"8","DOI":"10.1016\/j.infrared.2017.02.005","volume":"82","author":"J Ma","year":"2017","unstructured":"Ma, J., Zhou, Z., Wang, B., Zong, H.: Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys. Technol. 82, 8\u201317 (2017). https:\/\/doi.org\/10.1016\/j.infrared.2017.02.005. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1350449516305928","journal-title":"Infrared Phys. Technol."},{"key":"10_CR19","doi-asserted-by":"publisher","unstructured":"Mou, J., Gao, W., Song, Z.: Image fusion based on non-negative matrix factorization and infrared feature extraction. In: 2013 6th International Congress on Image and Signal Processing (CISP), vol.\u00a02, pp. 1046\u20131050 (2013). https:\/\/doi.org\/10.1109\/CISP.2013.6745210","DOI":"10.1109\/CISP.2013.6745210"},{"key":"10_CR20","doi-asserted-by":"publisher","unstructured":"Seif, G., Androutsos, D.: Edge-based loss function for single image super-resolution. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1468\u20131472 (2018). https:\/\/doi.org\/10.1109\/ICASSP.2018.8461664","DOI":"10.1109\/ICASSP.2018.8461664"},{"key":"10_CR21","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","volume":"82","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Ma, J.: Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network. Inform. Fusion 82, 28\u201342 (2022). https:\/\/doi.org\/10.1016\/j.inffus.2021.12.004. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1566253521002542","journal-title":"Inform. Fusion"},{"key":"10_CR22","doi-asserted-by":"publisher","first-page":"26119","DOI":"10.1109\/ACCESS.2024.3359715","volume":"12","author":"T Tong","year":"2024","unstructured":"Tong, T., Ye, A., Lu, Y., Wu, Z.: MDDCMA: a distributed image fusion framework based on multiscale dense dilated convolution and coordinate mean attention. IEEE Access 12, 26119\u201326136 (2024). https:\/\/doi.org\/10.1109\/ACCESS.2024.3359715","journal-title":"IEEE Access"},{"key":"10_CR23","doi-asserted-by":"publisher","first-page":"22190","DOI":"10.1109\/ACCESS.2024.3364050","volume":"12","author":"H Wang","year":"2024","unstructured":"Wang, H., Shu, C., Li, X., Fu, Y., Fu, Z., Yin, X.: Two-stream edge-aware network for infrared and visible image fusion with multi-level wavelet decomposition. IEEE Access 12, 22190\u201322204 (2024). https:\/\/doi.org\/10.1109\/ACCESS.2024.3364050","journal-title":"IEEE Access"},{"issue":"3","key":"10_CR24","doi-asserted-by":"publisher","first-page":"410","DOI":"10.1109\/JAS.2017.7510532","volume":"4","author":"W Wang","year":"2017","unstructured":"Wang, W., Yuan, X.: Recent advances in image dehazing. IEEE\/CAA Journal of Automatica Sinica 4(3), 410\u2013436 (2017). https:\/\/doi.org\/10.1109\/JAS.2017.7510532","journal-title":"IEEE\/CAA Journal of Automatica Sinica"},{"key":"10_CR25","doi-asserted-by":"publisher","unstructured":"Wei, R., Zhu, D., Zhan, W., Hao, Z.: Infrared and visible image fusion based on RPCA and NSST. In: 2019 IEEE International Conference on Power, Intelligent Computing and Systems (ICPICS), pp. 236\u2013240 (2019). https:\/\/doi.org\/10.1109\/ICPICS47731.2019.8942573","DOI":"10.1109\/ICPICS47731.2019.8942573"},{"issue":"1","key":"10_CR26","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2022","unstructured":"Xu, H., Ma, J., Jiang, J., Guo, X., Ling, H.: U2Fusion: a unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 502\u2013518 (2022). https:\/\/doi.org\/10.1109\/TPAMI.2020.3012548","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10_CR27","doi-asserted-by":"publisher","unstructured":"Xu, H., Ma, J., Le, Z., Jiang, J., Guo, X.: Fusiondn: a unified densely connected network for image fusion. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34(07), pp. 12484\u201312491, April 2020. https:\/\/doi.org\/10.1609\/aaai.v34i07.6936. https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/view\/6936","DOI":"10.1609\/aaai.v34i07.6936"},{"key":"10_CR28","doi-asserted-by":"publisher","first-page":"1134","DOI":"10.1109\/TCI.2021.3119954","volume":"7","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Yuan, J., Tian, X., Ma, J.: GAN-FM: Infrared and visible image fusion using GAN with full-scale skip connection and dual Markovian discriminators. IEEE Trans. Comput. Imaging 7, 1134\u20131147 (2021). https:\/\/doi.org\/10.1109\/TCI.2021.3119954","journal-title":"IEEE Trans. Comput. Imaging"},{"key":"10_CR29","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.infrared.2015.11.003","volume":"74","author":"Q Zhang","year":"2016","unstructured":"Zhang, Q., Maldague, X.: An adaptive fusion approach for infrared and visible images based on NSCT and compressed sensing. Infrared Phys. Technol. 74, 11\u201320 (2016). https:\/\/doi.org\/10.1016\/j.infrared.2015.11.003. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1350449515002662","journal-title":"Infrared Phys. Technol."},{"key":"10_CR30","doi-asserted-by":"publisher","unstructured":"Zhang, Y., Han, X., Zhang, H., Zhao, L.: Edge detection algorithm of image fusion based on improved sobel operator. In: 2017 IEEE 3rd Information Technology and Mechatronics Engineering Conference (ITOEC), pp. 457\u2013461 (2017). https:\/\/doi.org\/10.1109\/ITOEC.2017.8122336","DOI":"10.1109\/ITOEC.2017.8122336"},{"key":"10_CR31","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","volume":"54","author":"Y Zhang","year":"2020","unstructured":"Zhang, Y., Liu, Y., Sun, P., Yan, H., Zhao, X., Zhang, L.: IFCNN: a general image fusion framework based on convolutional neural network. Inform. Fusion 54, 99\u2013118 (2020). https:\/\/doi.org\/10.1016\/j.inffus.2019.07.011. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1566253518305505","journal-title":"Inform. Fusion"}],"container-title":["Lecture Notes in Computer Science","Knowledge Management and Acquisition for Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-0026-7_10","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,15]],"date-time":"2024-11-15T19:04:08Z","timestamp":1731697448000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-0026-7_10"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,16]]},"ISBN":["9789819600250","9789819600267"],"references-count":31,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-0026-7_10","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,16]]},"assertion":[{"value":"16 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PKAW","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Principle and Practice of Data and Knowledge Acquisition Workshop","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kyoto","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Japan","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 November 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"19 November 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"pkaw2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/pkawwebsite.github.io\/2024\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}