{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T00:09:52Z","timestamp":1773878992449,"version":"3.50.1"},"reference-count":59,"publisher":"Springer Science and Business Media LLC","issue":"15","license":[{"start":{"date-parts":[[2025,9,22]],"date-time":"2025-09-22T00:00:00Z","timestamp":1758499200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,9,22]],"date-time":"2025-09-22T00:00:00Z","timestamp":1758499200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Yunnan University Graduate Student Research and Innovation Fund Project Grant","award":["KC-23234062"],"award-info":[{"award-number":["KC-23234062"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62101481;62261060"],"award-info":[{"award-number":["62101481;62261060"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Yunnan Fundamental Research Projects","award":["202301AW070007;202201AU070033;202201AT070112;202301AU070210;202401AT070470"],"award-info":[{"award-number":["202301AW070007;202201AU070033;202201AT070112;202301AU070210;202401AT070470"]}]},{"name":"Major Scientific and Technological Project of Yunnan Province","award":["202202AD080002"],"award-info":[{"award-number":["202202AD080002"]}]},{"name":"Yunnan Province Expert Workstations","award":["202305AF150078"],"award-info":[{"award-number":["202305AF150078"]}]},{"name":"Xingdian Talent Project in Yunnan Province"},{"name":"National Science Center of Poland","award":["2020\/02\/Y\/ST6\/00037"],"award-info":[{"award-number":["2020\/02\/Y\/ST6\/00037"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s00371-025-04165-4","type":"journal-article","created":{"date-parts":[[2025,9,22]],"date-time":"2025-09-22T09:56:50Z","timestamp":1758535010000},"page":"12441-12459","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["DS-GAN: a dual sub-structure GAN for thermal infrared image colorization using U-Net with ConvNeXt and multi-scale large kernel attention"],"prefix":"10.1007","volume":"41","author":[{"given":"Guoliang","family":"Yao","sequence":"first","affiliation":[]},{"given":"Xin","family":"Jin","sequence":"additional","affiliation":[]},{"given":"Qian","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Michal","family":"Wozniak","sequence":"additional","affiliation":[]},{"given":"Shengfa","family":"Miao","sequence":"additional","affiliation":[]},{"given":"Shaowen","family":"Yao","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Zhou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,22]]},"reference":[{"key":"4165_CR1","doi-asserted-by":"crossref","unstructured":"Bhat, N., Saggu, N., Pragati, Kumar, S.: Generating visible spectrum images from thermal infrared using conditional generative adversarial networks. In: 2020 5th International Conference on Communication and Electronics Systems (ICCES), pp. 1390\u20131394 (2020)","DOI":"10.1109\/ICCES48766.2020.9137895"},{"key":"4165_CR2","doi-asserted-by":"crossref","unstructured":"Dong, Z., Qu, W.: Infrared image colorization using an edge aware auto encoder decoder with the multi-resolution fusion, In: Chinese Automation Congress (CAC). IEEE, vol. 2019, pp. 1011\u20131016 (2019)","DOI":"10.1109\/CAC48633.2019.8996588"},{"key":"4165_CR3","doi-asserted-by":"publisher","first-page":"245","DOI":"10.1007\/s00138-013-0570-5","volume":"25","author":"R Gade","year":"2014","unstructured":"Gade, R., Moeslund, T.B.: Thermal cameras and applications: a survey. Mach. Vis. Appl. 25, 245\u2013262 (2014)","journal-title":"Mach. Vis. Appl."},{"key":"4165_CR4","doi-asserted-by":"crossref","unstructured":"Zhang, S.,\u00a0Benenson, R.,\u00a0Schiele, B.: Citypersons: a diverse dataset for pedestrian detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3213\u20133221 (2017)","DOI":"10.1109\/CVPR.2017.474"},{"key":"4165_CR5","doi-asserted-by":"crossref","unstructured":"Zhang, S., Wen, L.,\u00a0Bian, X.,\u00a0Lei, Z., Li, S.\u00a0Z.: Occlusion-aware r-CNN: detecting pedestrians in a crowd. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 637\u2013653 (2018)","DOI":"10.1007\/978-3-030-01219-9_39"},{"issue":"1","key":"4165_CR6","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2020","unstructured":"Xu, H., Ma, J., Jiang, J., Guo, X., Ling, H.: U2Fusion: a unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 502\u2013518 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4165_CR7","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., Li, C., Jiang, J.: Fusiongan: a generative adversarial network for infrared and visible image fusion. Inf. Fus. 48, 11\u201326 (2019)","journal-title":"Inf. Fus."},{"key":"4165_CR8","doi-asserted-by":"publisher","first-page":"1707","DOI":"10.1007\/s00371-020-01933-2","volume":"37","author":"M Wu","year":"2021","unstructured":"Wu, M., Jin, X., Jiang, Q., Lee, S.-J., Liang, W., Lin, G., Yao, S.: Remote sensing image colorization using symmetrical multi-scale DCGAN in YUV color space. Vis. Comput. 37, 1707\u20131729 (2021)","journal-title":"Vis. Comput."},{"issue":"12","key":"4165_CR9","doi-asserted-by":"publisher","first-page":"6537","DOI":"10.1007\/s00371-022-02747-0","volume":"39","author":"K Xiong","year":"2023","unstructured":"Xiong, K., Hong, K., Li, J., Li, W., Liao, W., Liu, Q.: Joint intensity-gradient guided generative modeling for colorization. Vis. Comput. 39(12), 6537\u20136552 (2023)","journal-title":"Vis. Comput."},{"key":"4165_CR10","doi-asserted-by":"crossref","unstructured":"Berg, A.,\u00a0Ahlberg, J.,\u00a0Felsberg, M.: Generating visible spectrum images from thermal infrared. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 1143\u20131152 (2018)","DOI":"10.1109\/CVPRW.2018.00159"},{"key":"4165_CR11","unstructured":"Damer, N.,\u00a0Boutros, F.,\u00a0Mallat, K.,\u00a0Kirchbuchner, F., Dugelay, J.-L.,\u00a0Kuijper, A.: Cascaded generation of high-quality color visible face images from thermal captures (2019). [arXiv:1910.09524]"},{"key":"4165_CR12","doi-asserted-by":"publisher","first-page":"845","DOI":"10.1007\/s11263-019-01175-3","volume":"127","author":"H Zhang","year":"2019","unstructured":"Zhang, H., Riggan, B.S., Hu, S., Short, N.J., Patel, V.M.: Synthesis of high-quality visible faces from polarimetric thermal faces using generative adversarial networks. Int. J. Comput. Vis. 127, 845\u2013862 (2019)","journal-title":"Int. J. Comput. Vis."},{"key":"4165_CR13","doi-asserted-by":"publisher","DOI":"10.1016\/j.infrared.2020.103338","volume":"107","author":"X Kuang","year":"2020","unstructured":"Kuang, X., Zhu, J., Sui, X., Liu, Y., Liu, C., Chen, Q., Gu, G.: Thermal infrared colorization via conditional generative adversarial network. Infrared Phys. Technol. 107, 103338 (2020)","journal-title":"Infrared Phys. Technol."},{"issue":"8","key":"4165_CR14","doi-asserted-by":"publisher","first-page":"3062","DOI":"10.1109\/TCSVT.2020.3037688","volume":"31","author":"Y Zhao","year":"2020","unstructured":"Zhao, Y., Po, L.-M., Cheung, K.-W., Yu, W.-Y., Rehman, Y.A.U.: Scgan: saliency map-guided colorization with generative adversarial network. IEEE Trans. Circuits Syst. Video Technol. 31(8), 3062\u20133077 (2020)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"4","key":"4165_CR15","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2897824.2925974","volume":"35","author":"S Iizuka","year":"2016","unstructured":"Iizuka, S., Simo-Serra, E., Ishikawa, H.: Let there be color! joint end-to-end learning of global and local image priors for automatic image colorization with simultaneous classification. ACM Trans. Graph. (ToG) 35(4), 1\u201311 (2016)","journal-title":"ACM Trans. Graph. (ToG)"},{"key":"4165_CR16","doi-asserted-by":"crossref","unstructured":"Larsson, G., Maire, M., Shakhnarovich, G.: Learning representations for automatic colorization. In: Computer Vision-ECCV,: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, Part IV 14. Springer, vol. 2016, pp. 577\u2013593 (2016)","DOI":"10.1007\/978-3-319-46493-0_35"},{"key":"4165_CR17","doi-asserted-by":"crossref","unstructured":"Yang, B.,\u00a0Sun, F.,\u00a0Li, S.: Region-based color fusion method for visible and IR image sequences. In: 2008 Chinese Conference on Pattern Recognition. IEEE, pp. 1\u20136 (2008)","DOI":"10.1109\/CCPR.2008.34"},{"issue":"20","key":"4165_CR18","doi-asserted-by":"publisher","first-page":"6010","DOI":"10.1016\/j.ijleo.2014.07.059","volume":"125","author":"X Yu","year":"2014","unstructured":"Yu, X., Ren, J., Chen, Q., Sui, X.: A false color image fusion method based on multi-resolution color transfer in normalization ycbcr space. Optik 125(20), 6010\u20136016 (2014)","journal-title":"Optik"},{"key":"4165_CR19","doi-asserted-by":"crossref","unstructured":"Zhu, J.-Y.,\u00a0Park, T.,\u00a0Isola, P., Efros, A.\u00a0A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2223\u20132232 (2017)","DOI":"10.1109\/ICCV.2017.244"},{"issue":"9","key":"4165_CR20","doi-asserted-by":"publisher","first-page":"15808","DOI":"10.1109\/TITS.2022.3145476","volume":"23","author":"F Luo","year":"2022","unstructured":"Luo, F., Li, Y., Zeng, G., Peng, P., Wang, G., Li, Y.: Thermal infrared image colorization for nighttime driving scenes with top-down guided attention. IEEE Trans. Intell. Transp. Syst. 23(9), 15808\u201315823 (2022)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"4165_CR21","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.-Y.,\u00a0Zhou, T., Efros, A.\u00a0A.: Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1125\u20131134 (2017)","DOI":"10.1109\/CVPR.2017.632"},{"issue":"3","key":"4165_CR22","doi-asserted-by":"publisher","first-page":"478","DOI":"10.1109\/TCE.2023.3280165","volume":"69","author":"Y He","year":"2023","unstructured":"He, Y., Jin, X., Jiang, Q., Cheng, Z., Wang, P., Zhou, W.: LKAT-GAN: a GAN for thermal infrared image colorization based on large kernel and attentionunet-transformer. IEEE Trans. Consum. Electron. 69(3), 478\u2013489 (2023)","journal-title":"IEEE Trans. Consum. Electron."},{"issue":"4","key":"4165_CR23","doi-asserted-by":"publisher","first-page":"2954","DOI":"10.1109\/TIV.2022.3218833","volume":"8","author":"H Liao","year":"2022","unstructured":"Liao, H., Jiang, Q., Jin, X., Liu, L., Liu, L., Lee, S.-J., Zhou, W.: Mugan: thermal infrared image colorization using mixed-skipping unet and generative adversarial network. IEEE Trans. Intell. Veh. 8(4), 2954\u20132969 (2022)","journal-title":"IEEE Trans. Intell. Veh."},{"key":"4165_CR24","doi-asserted-by":"crossref","unstructured":"Liu, Z.,\u00a0Mao, H., Wu, C.-Y.,\u00a0Feichtenhofer, C.,\u00a0Darrell, T.,\u00a0Xie, S.: A convnet for the 2020s. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11976\u201311986 (2022)","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"4165_CR25","unstructured":"Howard, A.\u00a0G.,\u00a0Zhu, M.,\u00a0Chen, B.,\u00a0Kalenichenko, D.,\u00a0Wang, W.,\u00a0Weyand, T.,\u00a0Andreetto, M.,\u00a0Adam, H.: Mobilenets: efficient convolutional neural networks for mobile vision applications (2017). [arXiv:1704.04861]"},{"key":"4165_CR26","doi-asserted-by":"crossref","unstructured":"Chollet, F.: Xception: Deep learning with depthwise separable convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1251\u20131258 (2017)","DOI":"10.1109\/CVPR.2017.195"},{"key":"4165_CR27","unstructured":"Nair, V., Hinton, G.\u00a0E.: Rectified linear units improve restricted boltzmann machines. In: Proceedings of the 27th International Conference on Machine Learning (ICML-10), pp. 807\u2013814 (2010)"},{"key":"4165_CR28","unstructured":"Hendrycks, D.,\u00a0Gimpel, K.: Gaussian error linear units (gelus) (2016). [arXiv:1606.08415]"},{"key":"4165_CR29","unstructured":"Ba, J.\u00a0L., Kiros, J.\u00a0R., Hinton, G.\u00a0E.: Layer normalization (2016). [arXiv:1607.06450]"},{"key":"4165_CR30","unstructured":"Ulyanov, D.,\u00a0Vedaldi, A.,\u00a0Lempitsky, V.: Instance normalization: the missing ingredient for fast stylization (2016). [arXiv:1607.08022]"},{"key":"4165_CR31","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention-MICCAI: 18th International Conference, Munich, Germany, October 5\u20139, 2015, Proceedings, Part III 18, vol. 2015, pp. 234\u2013241. Springer (2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"issue":"5","key":"4165_CR32","doi-asserted-by":"publisher","first-page":"1120","DOI":"10.1109\/TIP.2005.864231","volume":"15","author":"L Yatziv","year":"2006","unstructured":"Yatziv, L., Sapiro, G.: Fast image and video colorization using chrominance blending. IEEE Trans. Image Process. 15(5), 1120\u20131129 (2006)","journal-title":"IEEE Trans. Image Process."},{"key":"4165_CR33","doi-asserted-by":"crossref","unstructured":"S\u1ef3kora, D., Dingliana, J., Collins, S.: Lazybrush: Flexible painting tool for hand-drawn cartoons. In: Computer Graphics Forum, Wiley Online Library, Vol. 28, No. 2, pp. 599\u2013608 (2009)","DOI":"10.1111\/j.1467-8659.2009.01400.x"},{"key":"4165_CR34","doi-asserted-by":"crossref","unstructured":"Sangkloy, P.,\u00a0Lu, J.,\u00a0Fang, C.,\u00a0Yu, F.,\u00a0Hays, J.: Scribbler: controlling deep image synthesis with sketch and color. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5400\u20135409 (2017)","DOI":"10.1109\/CVPR.2017.723"},{"key":"4165_CR35","doi-asserted-by":"crossref","unstructured":"Pang, J. Au, O.\u00a0C.,\u00a0Yamashita, Y.,\u00a0Ling, Y.,\u00a0Guo, Y.,\u00a0Zeng, J.: Self-similarity-based image colorization. In: 2014 IEEE International Conference on Image Processing (ICIP). IEEE, pp. 4687\u20134691 (2014)","DOI":"10.1109\/ICIP.2014.7025950"},{"key":"4165_CR36","doi-asserted-by":"publisher","first-page":"21604","DOI":"10.1109\/ACCESS.2021.3055575","volume":"9","author":"K Du","year":"2021","unstructured":"Du, K., Liu, C., Cao, L., Guo, Y., Zhang, F., Wang, T.: Double-channel guided generative adversarial network for image colorization. IEEE Access 9, 21604\u201321617 (2021)","journal-title":"IEEE Access"},{"key":"4165_CR37","doi-asserted-by":"publisher","first-page":"8526","DOI":"10.1109\/TIP.2021.3117061","volume":"30","author":"H Li","year":"2021","unstructured":"Li, H., Sheng, B., Li, P., Ali, R., Chen, C.P.: Globally and locally semantic colorization via exemplar-based broad-GAN. IEEE Trans. Image Process. 30, 8526\u20138539 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"4165_CR38","doi-asserted-by":"crossref","unstructured":"Su, J.-W., Chu, H.-K., Huang, J.-B.: Instance-aware image colorization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7968\u20137977 (2020)","DOI":"10.1109\/CVPR42600.2020.00799"},{"key":"4165_CR39","doi-asserted-by":"crossref","unstructured":"Cheng, Z.,\u00a0Yang, Q.,\u00a0Sheng, B.: Deep colorization. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 415\u2013423 (2015)","DOI":"10.1109\/ICCV.2015.55"},{"key":"4165_CR40","doi-asserted-by":"publisher","first-page":"818","DOI":"10.1007\/s11263-019-01271-4","volume":"128","author":"J Zhao","year":"2020","unstructured":"Zhao, J., Han, J., Shao, L., Snoek, C.G.: Pixelated semantic colorization. Int. J. Comput. Vis. 128, 818\u2013834 (2020)","journal-title":"Int. J. Comput. Vis."},{"key":"4165_CR41","unstructured":"Zhao, J.,\u00a0Liu, L., Snoek, C.\u00a0G.,\u00a0Han, J.,\u00a0Shao, L.: Pixel-level semantics guided image colorization (2018). [arXiv:1808.01597]"},{"issue":"11","key":"4165_CR42","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1145\/3422622","volume":"63","author":"I Goodfellow","year":"2020","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial networks. Commun. ACM 63(11), 139\u2013144 (2020)","journal-title":"Commun. ACM"},{"key":"4165_CR43","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A.: Colorful image colorization. In: Computer Vision-ECCV: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, Part III 14. Springer 2016, pp. 649\u2013666 (2016)","DOI":"10.1007\/978-3-319-46487-9_40"},{"issue":"17","key":"4165_CR44","doi-asserted-by":"publisher","first-page":"26465","DOI":"10.1007\/s11042-021-10881-5","volume":"80","author":"S Huang","year":"2021","unstructured":"Huang, S., Jin, X., Jiang, Q., Li, J., Lee, S.-J., Wang, P., Yao, S.: A fully-automatic image colorization scheme using improved Cyclegan with skip connections. Multimedia Tools Appl. 80(17), 26465\u201326492 (2021)","journal-title":"Multimedia Tools Appl."},{"key":"4165_CR45","first-page":"2672","volume":"27","author":"I Goodfellow","year":"2014","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. Adv. Neural. Inf. Process. Syst. 27, 2672\u20132680 (2014)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"4165_CR46","doi-asserted-by":"crossref","unstructured":"Li, S.,\u00a0Han, B.,\u00a0Yu, Z., Liu, C.\u00a0H.,\u00a0Chen, K.,\u00a0Wang, S.: I2v-gan: unpaired infrared-to-visible video translation. In: Proceedings of the 29th ACM International Conference on Multimedia, pp. 3061\u20133069 (2021)","DOI":"10.1145\/3474085.3475445"},{"key":"4165_CR47","doi-asserted-by":"crossref","unstructured":"Bansal, A.,\u00a0Ma, S.,\u00a0Ramanan, D.,\u00a0Sheikh, Y.: Recycle-gan: unsupervised video retargeting. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 119\u2013135 (2018)","DOI":"10.1007\/978-3-030-01228-1_8"},{"key":"4165_CR48","doi-asserted-by":"crossref","unstructured":"Anoosheh, A.,\u00a0Sattler, T.,\u00a0Timofte, R.,\u00a0Pollefeys, M.,\u00a0Van\u00a0Gool, L.: Night-to-day image translation for retrieval-based localization. In: 2019 International Conference on Robotics and Automation (ICRA). IEEE, pp. 5958\u20135964 (2019)","DOI":"10.1109\/ICRA.2019.8794387"},{"key":"4165_CR49","doi-asserted-by":"crossref","unstructured":"Huang, H.,\u00a0Lin, L.,\u00a0Tong, R.,\u00a0Hu, H.,\u00a0Zhang, Q.,\u00a0Iwamoto, Y.,\u00a0Han, X., Chen, Y.-W.,\u00a0Wu, J.: Unet 3+: A full-scale connected unet for medical image segmentation. In: ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, pp. 1055\u20131059 (2020)","DOI":"10.1109\/ICASSP40776.2020.9053405"},{"key":"4165_CR50","doi-asserted-by":"publisher","first-page":"318","DOI":"10.1016\/j.neucom.2022.06.021","volume":"501","author":"H Wang","year":"2022","unstructured":"Wang, H., Cheng, C., Zhang, X., Sun, H.: Towards high-quality thermal infrared image colorization via attention-based hierarchical network. Neurocomputing 501, 318\u2013327 (2022)","journal-title":"Neurocomputing"},{"key":"4165_CR51","unstructured":"Chen, L.-C.,\u00a0Papandreou, G.,\u00a0Schroff, F.,\u00a0Adam, H.: Rethinking atrous convolution for semantic image segmentation (2017). [arXiv:1706.05587]"},{"key":"4165_CR52","doi-asserted-by":"crossref","unstructured":"Woo, S.,\u00a0Park, J., Lee, J.-Y., Kweon, I.\u00a0S.: Cbam: convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"4165_CR53","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., et al.: Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 115, 211\u2013252 (2015)","journal-title":"Int. J. Comput. Vis."},{"key":"4165_CR54","doi-asserted-by":"crossref","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","DOI":"10.1109\/TIP.2003.819861"},{"key":"4165_CR55","doi-asserted-by":"crossref","unstructured":"Hwang, S.,\u00a0Park, J.,\u00a0Kim, N.,\u00a0Choi, Y.,\u00a0So\u00a0Kweon, I.: Multispectral pedestrian detection: benchmark dataset and baseline. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1037\u20131045 (2015)","DOI":"10.1109\/CVPR.2015.7298706"},{"key":"4165_CR56","unstructured":"Paszke, A.,\u00a0Gross, S.,\u00a0Massa, F.,\u00a0Lerer, A.,\u00a0Bradbury, J.,\u00a0Chanan, G.,\u00a0Killeen, T.,\u00a0Lin, Z.,\u00a0Gimelshein, N.,\u00a0Antiga L., et\u00a0al.: Pytorch: an imperative style, high-performance deep learning library. In: Advances in Neural Information Processing Systems, vol. 32 (2019)"},{"key":"4165_CR57","unstructured":"Kingma, D.\u00a0P.,\u00a0Ba, J.: Adam: a method for stochastic optimization (2014). [arXiv:1412.6980]"},{"key":"4165_CR58","unstructured":"Kodali, N.,\u00a0Abernethy, J.,\u00a0Hays, J.,\u00a0Kira, Z.: On convergence and stability of gans (2017). [arXiv:1705.07215]"},{"key":"4165_CR59","unstructured":"Odena, A.,\u00a0Buckman, J.,\u00a0Olsson, C.,\u00a0Brown, T.,\u00a0Olah, C.,\u00a0Raffel, C.,\u00a0Goodfellow, I.: \u201cIs generator conditioning causally related to GAN performance? In: International Conference on Machine Learning. PMLR, pp. 3849\u20133858 (2018)"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04165-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04165-4\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04165-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T13:17:01Z","timestamp":1763644621000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04165-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,22]]},"references-count":59,"journal-issue":{"issue":"15","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["4165"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04165-4","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,9,22]]},"assertion":[{"value":"13 July 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 August 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 September 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"The authors declare no Conflict of interest.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}