{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,25]],"date-time":"2025-09-25T00:14:50Z","timestamp":1758759290242,"version":"3.44.0"},"reference-count":70,"publisher":"Springer Science and Business Media LLC","issue":"13","license":[{"start":{"date-parts":[[2025,7,23]],"date-time":"2025-07-23T00:00:00Z","timestamp":1753228800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,7,23]],"date-time":"2025-07-23T00:00:00Z","timestamp":1753228800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52273228"],"award-info":[{"award-number":["52273228"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Key Program of Science and Technology of Yunnan Province","award":["202302AB080022"],"award-info":[{"award-number":["202302AB080022"]}]},{"name":"the Project of Key Laboratory of Silicate Cultural Relics Conservation (Shanghai University), Ministry of Education","award":["SCRC2023ZZ07TS"],"award-info":[{"award-number":["SCRC2023ZZ07TS"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,10]]},"DOI":"10.1007\/s00371-025-04104-3","type":"journal-article","created":{"date-parts":[[2025,7,23]],"date-time":"2025-07-23T11:08:24Z","timestamp":1753268904000},"page":"11239-11259","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Geodesic feature augmentation for zero-shot text-guided diffusion style transfer"],"prefix":"10.1007","volume":"41","author":[{"given":"Yuexing","family":"Han","sequence":"first","affiliation":[]},{"given":"Liheng","family":"Ruan","sequence":"additional","affiliation":[]},{"given":"Bing","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,23]]},"reference":[{"key":"4104_CR1","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., Polosukhin, I.: Attention is all you need. Adv. Neural Inf. Process. Syst. 30 (2017)"},{"key":"4104_CR2","unstructured":"OpenAI: Gpt-4 technical report (2024). arXiv:2303.08774"},{"key":"4104_CR3","unstructured":"Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., et\u00a0al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning. PMLR, pp. 8748\u20138763 (2021)"},{"key":"4104_CR4","unstructured":"Li, J., Li, D., Xiong, C., Hoi, S.: Blip: bootstrapping language-image pre-training for unified vision-language understanding and generation. In: International Conference on Machine Learning. PMLR, pp. 12888\u201312900 (2022)"},{"key":"4104_CR5","doi-asserted-by":"crossref","unstructured":"Patashnik, O., Wu, Z., Shechtman, E., Cohen-Or, D., Lischinski, D.: StyleCLIP: text-driven manipulation of StyleGAN imagery. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 2085\u20132094 (2021)","DOI":"10.1109\/ICCV48922.2021.00209"},{"issue":"4","key":"4104_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3528223.3530164","volume":"41","author":"R Gal","year":"2022","unstructured":"Gal, R., Patashnik, O., Maron, H., Bermano, A.H., Chechik, G., Cohen-Or, D.: Stylegan-nada: clip-guided domain adaptation of image generators. ACM Trans. Graph. (TOG) 41(4), 1\u201313 (2022)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"4104_CR7","doi-asserted-by":"crossref","unstructured":"Crowson, K., Biderman, S., Kornis, D., Stander, D., Hallahan, E., Castricato, L., Raff, E.: Vqgan-clip: open domain image generation and editing with natural language guidance. In: European Conference on Computer Vision. Springer, pp. 88\u2013105 (2022)","DOI":"10.1007\/978-3-031-19836-6_6"},{"key":"4104_CR8","doi-asserted-by":"crossref","unstructured":"Kwon, G., Ye, J.C.: Clipstyler: image style transfer with a single text condition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2022), pp. 18062\u201318071","DOI":"10.1109\/CVPR52688.2022.01753"},{"key":"4104_CR9","first-page":"6840","volume":"33","author":"J Ho","year":"2020","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. Adv. Neural. Inf. Process. Syst. 33, 6840\u20136851 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"4104_CR10","first-page":"8780","volume":"34","author":"P Dhariwal","year":"2021","unstructured":"Dhariwal, P., Nichol, A.: Diffusion models beat gans on image synthesis. Adv. Neural. Inf. Process. Syst. 34, 8780\u20138794 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"4104_CR11","first-page":"36479","volume":"35","author":"C Saharia","year":"2022","unstructured":"Saharia, C., Chan, W., Saxena, S., Li, L., Whang, J., Denton, E.L., Ghasemipour, K., Gontijo Lopes, R., Karagol Ayan, B., Salimans, T., et al.: Photorealistic text-to-image diffusion models with deep language understanding. Adv. Neural. Inf. Process. Syst. 35, 36479\u201336494 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"4104_CR12","unstructured":"Ramesh, A., Dhariwal, P., Nichol, A., Chu, C., Chen, M.: Hierarchical text-conditional image generation with clip latents. 1(2), 3. arXiv preprint arXiv:2204.06125 (2022)"},{"key":"4104_CR13","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10684\u201310695 (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"4104_CR14","doi-asserted-by":"crossref","unstructured":"Dong, W., Xue, S., Duan, X., Han, S.: Prompt tuning inversion for text-driven image editing using diffusion models. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 7430\u20137440 (2023)","DOI":"10.1109\/ICCV51070.2023.00683"},{"key":"4104_CR15","doi-asserted-by":"crossref","unstructured":"Xu, S., Huang, Y., Pan, J., Ma, Z., Chai, J.: Inversion-free image editing with language-guided diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9452\u20139461 (2024)","DOI":"10.1109\/CVPR52733.2024.00903"},{"key":"4104_CR16","unstructured":"Podell, D., English, Z., Lacey, K., Blattmann, A., Dockhorn, T., M\u00fcller, J., Penna, J., Rombach, R.: SDXL: improving latent diffusion models for high-resolution image synthesis. In: The Twelfth International Conference on Learning Representations (2024). https:\/\/openreview.net\/forum?id=di52zR8xgf"},{"key":"4104_CR17","unstructured":"Ju, X., Zeng, A., Bian, Y., Liu, S., Xu, Q.: PnP inversion: boosting diffusion-based editing with 3 lines of code. In: The Twelfth International Conference on Learning Representations (2024). https:\/\/openreview.net\/forum?id=FoMZ4ljhVw"},{"key":"4104_CR18","unstructured":"Esser, P., Kulal, S., Blattmann, A., Entezari, R., M\u00fcller, J., Saini, H., Levi, Y., Lorenz, D., Sauer, A., Boesel, F., Podell, D., Dockhorn, T., English, Z., Rombach, R.: Scaling rectified flow transformers for high-resolution image synthesis. In: Proceedings of the 41st International Conference on Machine Learning (JMLR.org, 2024), ICML\u201924"},{"key":"4104_CR19","unstructured":"Labs, B.F.: Flux. https:\/\/github.com\/black-forest-labs\/flux (2024)"},{"key":"4104_CR20","unstructured":"Su, X., Song, J., Meng, C., Ermon, S.: Dual diffusion implicit bridges for image-to-image translation. In: The Eleventh International Conference on Learning Representations (2022)"},{"key":"4104_CR21","doi-asserted-by":"crossref","unstructured":"Wang, Z., Zhao, L., Xing, W.: Stylediffusion: controllable disentangled style transfer via diffusion models. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 7677\u20137689 (2023)","DOI":"10.1109\/ICCV51070.2023.00706"},{"key":"4104_CR22","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Huang, N., Tang, F., Huang, H., Ma, C., Dong, W., Xu, C.: Inversion-based style transfer with diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10146\u201310156 (2023)","DOI":"10.1109\/CVPR52729.2023.00978"},{"key":"4104_CR23","doi-asserted-by":"crossref","unstructured":"Cho, H., Lee, J., Chang, S., Jeong, Y.: One-shot structure-aware stylized image synthesis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2024), pp. 8302\u20138311","DOI":"10.1109\/CVPR52733.2024.00793"},{"key":"4104_CR24","doi-asserted-by":"crossref","unstructured":"Chung, J., Hyun, S., Heo, J.P.: Style injection in diffusion: a training-free approach for adapting large-scale diffusion models for style transfer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8795\u20138805 (2024)","DOI":"10.1109\/CVPR52733.2024.00840"},{"key":"4104_CR25","doi-asserted-by":"crossref","unstructured":"Deng, Y., He, X., Tang, F., Dong, W.: Z*: zero-shot style transfer via attention reweighting. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2024), pp. 6934\u20136944","DOI":"10.1109\/CVPR52733.2024.00662"},{"key":"4104_CR26","unstructured":"Song, J., Meng, C., Ermon, S.: Denoising diffusion implicit models. In: International Conference on Learning Representations (2021). https:\/\/openreview.net\/forum?id=St1giarCHLP"},{"key":"4104_CR27","unstructured":"Kwon, G., Ye, J.C.: Diffusion-based image translation using disentangled style and content representation. In: The Eleventh International Conference on Learning Representations (2023). https:\/\/openreview.net\/forum?id=Nayau9fwXU"},{"key":"4104_CR28","doi-asserted-by":"crossref","unstructured":"Yang, S., Hwang, H., Ye, J.C.: Zero-shot contrastive loss for text-guided diffusion image style transfer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (2023), pp. 22873\u201322882","DOI":"10.1109\/ICCV51070.2023.02091"},{"key":"4104_CR29","unstructured":"He, F., Li, G., Zhang, M., Yan, L., Si, L., Li, F.: Freestyle: free lunch for text-guided style transfer using diffusion models. arXiv preprint arXiv:2401.15636 (2024)"},{"key":"4104_CR30","doi-asserted-by":"publisher","first-page":"49","DOI":"10.1023\/A:1026553619983","volume":"40","author":"J Portilla","year":"2000","unstructured":"Portilla, J., Simoncelli, E.P.: A parametric texture model based on joint statistics of complex wavelet coefficients. Int. J. Comput. Vis. 40, 49\u201370 (2000)","journal-title":"Int. J. Comput. Vis."},{"key":"4104_CR31","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. Adv. Neural Inf. Process. Syst. 27 (2014)"},{"key":"4104_CR32","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision (2017), pp. 2223\u20132232","DOI":"10.1109\/ICCV.2017.244"},{"issue":"4","key":"4104_CR33","doi-asserted-by":"publisher","first-page":"1083","DOI":"10.1109\/TMI.2022.3223683","volume":"42","author":"R Liu","year":"2022","unstructured":"Liu, R., Wang, T., Li, H., Zhang, P., Li, J., Yang, X., Shen, D., Sheng, B.: TMM-Nets: transferred multi-to mono-modal generation for lupus retinopathy diagnosis. IEEE Trans. Med. Imaging 42(4), 1083\u20131094 (2022)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"4104_CR34","doi-asserted-by":"crossref","unstructured":"Park, T., Efros, A.A., Zhang, R., Zhu, J.Y.: Contrastive learning for unpaired image-to-image translation. In: Computer Vision-ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part IX 16. Springer, pp. 319\u2013345 (2020)","DOI":"10.1007\/978-3-030-58545-7_19"},{"issue":"3","key":"4104_CR35","doi-asserted-by":"publisher","first-page":"e2248","DOI":"10.1002\/cav.2248","volume":"35","author":"W Zhao","year":"2024","unstructured":"Zhao, W., Zhu, J., Huang, J., Li, P., Sheng, B.: GAN-based multi-decomposition photo cartoonization. Comput. Anim. Virtual Worlds 35(3), e2248 (2024)","journal-title":"Comput. Anim. Virtual Worlds"},{"key":"4104_CR36","doi-asserted-by":"crossref","unstructured":"Huang, X., Belongie, S.: Arbitrary style transfer in real-time with adaptive instance normalization. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1501\u20131510 (2017)","DOI":"10.1109\/ICCV.2017.167"},{"key":"4104_CR37","doi-asserted-by":"crossref","unstructured":"Liu, S., Lin, T., He, D., Li, F., Wang, M., Li, X., Sun, Z., Li, Q., Ding, E.: Adaattn: revisit attention mechanism in arbitrary neural style transfer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6649\u20136658 (2021)","DOI":"10.1109\/ICCV48922.2021.00658"},{"issue":"3","key":"4104_CR38","doi-asserted-by":"publisher","first-page":"1369","DOI":"10.1007\/s00371-023-02855-5","volume":"40","author":"X Yu","year":"2024","unstructured":"Yu, X., Zhou, G.: Arbitrary style transfer via content consistency and style consistency. Vis. Comput. 40(3), 1369\u20131382 (2024)","journal-title":"Vis. Comput."},{"key":"4104_CR39","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Tang, F., Dong, W., Huang, H., Ma, C., Lee, T.Y., Xu, C.: Domain enhanced arbitrary image style transfer via contrastive learning. In: ACM SIGGRAPH 2022 Conference Proceedings, pp. 1\u20138 (2022)","DOI":"10.1145\/3528233.3530736"},{"key":"4104_CR40","doi-asserted-by":"crossref","unstructured":"You, C., Zhao, R., Staib, L.H., Duncan, J.S.: Momentum contrastive voxel-wise representation learning for semi-supervised volumetric medical image segmentation. In: Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S. (eds.) Medical image computing and computer assisted intervention\u2014MICCAI 2022. Springer, Cham (2022), pp. 639\u2013652","DOI":"10.1007\/978-3-031-16440-8_61"},{"issue":"9","key":"4104_CR41","doi-asserted-by":"publisher","first-page":"2228","DOI":"10.1109\/TMI.2022.3161829","volume":"41","author":"C You","year":"2022","unstructured":"You, C., Zhou, Y., Zhao, R., Staib, L., Duncan, J.S.: Simcvd: simple contrastive voxel-wise representation distillation for semi-supervised medical image segmentation. IEEE Trans. Med. Imaging 41(9), 2228\u20132237 (2022)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"4104_CR42","doi-asserted-by":"crossref","unstructured":"You, C., Dai, W., Min, Y., Staib, L., Duncan, J.S.: Bootstrapping semi-supervised medical image segmentation with anatomical-aware contrastive distillation. In: International Conference on Information Processing in Medical Imaging. Springer, pp. 641\u2013653 (2023)","DOI":"10.1007\/978-3-031-34048-2_49"},{"key":"4104_CR43","doi-asserted-by":"crossref","unstructured":"You, C., Dai, W., Min, Y., Staib, L., Sekhon, J., Duncan, J.S.: Action++: improving semi-supervised medical image segmentation with adaptive anatomical contrast. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, pp. 194\u2013205 (2023)","DOI":"10.1007\/978-3-031-43901-8_19"},{"issue":"12","key":"4104_CR44","doi-asserted-by":"publisher","first-page":"11136","DOI":"10.1109\/TPAMI.2024.3461321","volume":"46","author":"C You","year":"2024","unstructured":"You, C., Dai, W., Liu, F., Min, Y., Dvornek, N.C., Li, X., Clifton, D.A., Staib, L., Duncan, J.S.: Mine yOur owN anatomy: revisiting medical image segmentation with extremely limited labels. IEEE Trans. Pattern Anal. Mach. Intell. 46(12), 11136\u201311151 (2024). https:\/\/doi.org\/10.1109\/TPAMI.2024.3461321","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4104_CR45","first-page":"9984","volume":"36","author":"C You","year":"2023","unstructured":"You, C., Dai, W., Min, Y., Liu, F., Clifton, D., Zhou, S.K., Staib, L., Duncan, J.: Rethinking semi-supervised medical image segmentation: a variance-reduction perspective. Adv. Neural. Inf. Process. Syst. 36, 9984\u201310021 (2023)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"4104_CR46","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et\u00a0al.: An image is worth 16 $$\\times $$ 16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"4104_CR47","first-page":"29582","volume":"35","author":"C You","year":"2022","unstructured":"You, C., Zhao, R., Liu, F., Dong, S., Chinchali, S., Topcu, U., Staib, L., Duncan, J.: Class-aware adversarial transformers for medical image segmentation. Adv. Neural. Inf. Process. Syst. 35, 29582\u201329596 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"4104_CR48","doi-asserted-by":"crossref","unstructured":"Deng, Y., Tang, F., Dong, W., Ma, C., Pan, X., Wang, L., Xu, C.: Stytr2: image style transfer with transformers. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11326\u201311336 (2022)","DOI":"10.1109\/CVPR52688.2022.01104"},{"key":"4104_CR49","unstructured":"Sohl-Dickstein, J., Weiss, E., Maheswaranathan, N., Ganguli, S.: Deep unsupervised learning using nonequilibrium thermodynamics. In: International Conference on Machine Learning. PMLR, pp. 2256\u20132265 (2015)"},{"issue":"1","key":"4104_CR50","doi-asserted-by":"publisher","first-page":"394","DOI":"10.1109\/TIP.2018.2866698","volume":"28","author":"WR Tan","year":"2019","unstructured":"Tan, W.R., Chan, C.S., Aguirre, H., Tanaka, K.: Improved artgan for conditional synthesis of natural image and artwork. IEEE Trans. Image Process. 28(1), 394\u2013409 (2019). https:\/\/doi.org\/10.1109\/TIP.2018.2866698","journal-title":"IEEE Trans. Image Process."},{"key":"4104_CR51","unstructured":"Face. H.: Using diffusers for image-to-image tasks (2024). https:\/\/huggingface.co\/docs\/diffusers\/en\/using-diffusers\/img2img. Accessed 15 July 2024"},{"issue":"2","key":"4104_CR52","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1112\/blms\/16.2.81","volume":"16","author":"DG Kendall","year":"1984","unstructured":"Kendall, D.G.: Shape manifolds, procrustean metrics, and complex projective spaces. Bull. Lond. Math. Soc. 16(2), 81\u2013121 (1984)","journal-title":"Bull. Lond. Math. Soc."},{"issue":"3","key":"4104_CR53","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/1276377.1276457","volume":"26","author":"M Kilian","year":"2007","unstructured":"Kilian, M., Mitra, N.J., Pottmann, H.: Geometric modeling in shape space. ACM Trans. Graph. (SIGGRAPH) 26(3), 1\u20138 (2007)","journal-title":"ACM Trans. Graph. (SIGGRAPH)"},{"issue":"4","key":"4104_CR54","doi-asserted-by":"publisher","first-page":"1467","DOI":"10.1016\/j.patcog.2009.09.028","volume":"43","author":"Y Han","year":"2010","unstructured":"Han, Y., Wang, B., Idesawa, M., Shimai, H.: Recognition of multiple configurations of objects with limited data. Pattern Recogn. 43(4), 1467\u20131475 (2010)","journal-title":"Pattern Recogn."},{"key":"4104_CR55","doi-asserted-by":"publisher","first-page":"195","DOI":"10.1007\/s10044-012-0277-7","volume":"17","author":"Y Han","year":"2014","unstructured":"Han, Y., Koike, H., Idesawa, M.: Recognizing objects with multiple configurations. Pattern Anal. Appl. 17, 195\u2013209 (2014)","journal-title":"Pattern Anal. Appl."},{"key":"4104_CR56","doi-asserted-by":"crossref","unstructured":"Paskin, M., Baum, D., Dean, M.N., von Tycowicz, C.: A Kendall shape space approach to 3D shape estimation from 2D landmarks. In: European Conference on Computer Vision. Springer, pp. 363\u2013379 (2022)","DOI":"10.1007\/978-3-031-20086-1_21"},{"key":"4104_CR57","doi-asserted-by":"crossref","unstructured":"Friji, R., Drira, H., Chaieb, F., Kchok, H., Kurtek, S.: Geometric deep neural network using rigid and non-rigid transformations for human action recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 12611\u201312620 (2021)","DOI":"10.1109\/ICCV48922.2021.01238"},{"key":"4104_CR58","unstructured":"Han, Y., Wan, G., Wang, B.: FAGC: feature augmentation on geodesic curve in the pre-shape space (2023). arXiv:2312.03325"},{"key":"4104_CR59","unstructured":"Han, Y., Ruan, L., Wang, B.: Few-shot image generation via information transfer from the built geodesic surface (2024). arXiv:2401.01749"},{"key":"4104_CR60","doi-asserted-by":"crossref","unstructured":"You, C., Dai, W., Min, Y., Staib, L., Duncan, J.S.: Implicit anatomical rendering for medical image segmentation with stochastic experts. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, pp. 561\u2013571 (2023)","DOI":"10.1007\/978-3-031-43898-1_54"},{"issue":"8","key":"4104_CR61","doi-asserted-by":"publisher","first-page":"4499","DOI":"10.1109\/TNNLS.2021.3116209","volume":"34","author":"Z Xie","year":"2021","unstructured":"Xie, Z., Zhang, W., Sheng, B., Li, P., Chen, C.P.: BAGFN: broad attentive graph fusion network for high-order feature interactions. IEEE Trans. Neural Netw. Learn. Syst. 34(8), 4499\u20134513 (2021)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"4104_CR62","doi-asserted-by":"crossref","unstructured":"Xiao, J., Li, L., Wang, C., Zha, Z.J., Huang, Q.: Few shot generative model adaption via relaxed spatial structural alignment. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11204\u201311213 (2022)","DOI":"10.1109\/CVPR52688.2022.01092"},{"key":"4104_CR63","doi-asserted-by":"crossref","unstructured":"You, C., Xiang, J., Su, K., Zhang, X., Dong, S., Onofrey, J., Staib, L., Duncan, J.S.: Incremental learning meets transfer learning: application to multi-site prostate MRI segmentation. In: International Workshop on Distributed, Collaborative, and Federated Learning. Springer, pp. 3\u201316 (2022)","DOI":"10.1007\/978-3-031-18523-6_1"},{"key":"4104_CR64","doi-asserted-by":"crossref","unstructured":"You, C., Yang, J., Chapiro, J., Duncan, J.S.: Unsupervised Wasserstein distance guided domain adaptation for 3D multi-domain liver segmentation. In: Interpretable and Annotation-Efficient Learning for Medical Image Computing: Third International Workshop, IMIMIC 2020, Second International Workshop, MIL3ID 2020, and 5th International Workshop, LABELS 2020, Held in Conjunction with MICCAI 2020, Lima, Peru, October 4\u20138, 2020, Proceedings 3. Springer, pp. 155\u2013163 (2020)","DOI":"10.1007\/978-3-030-61166-8_17"},{"key":"4104_CR65","unstructured":"Baranchuk, D., Voynov, A., Rubachev, I., Khrulkov, V., Babenko, A.: Label-efficient semantic segmentation with diffusion models. In: International Conference on Learning Representations (2021)"},{"key":"4104_CR66","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 28 (2015)"},{"key":"4104_CR67","doi-asserted-by":"publisher","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255 (2009). https:\/\/doi.org\/10.1109\/CVPR.2009.5206848","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"4104_CR68","doi-asserted-by":"publisher","unstructured":"Hor\u00e9, A., Ziou, D.: Image quality metrics: PSNR vs. SSIM. In: 2010 20th International Conference on Pattern Recognition, pp. 2366\u20132369 (2010). https:\/\/doi.org\/10.1109\/ICPR.2010.579","DOI":"10.1109\/ICPR.2010.579"},{"key":"4104_CR69","doi-asserted-by":"publisher","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition. IEEE, Salt Lake City, pp. 586\u2013595 (2018). https:\/\/doi.org\/10.1109\/CVPR.2018.00068. https:\/\/ieeexplore.ieee.org\/document\/8578166\/","DOI":"10.1109\/CVPR.2018.00068"},{"key":"4104_CR70","unstructured":"von Platen, P., Patil, S., Lozhkov, A., Cuenca, P., Lambert, N., Rasul, K., Davaadorj, M., Nair, D., Paul, S., Berman, W., Xu, Y., Liu, S., Wolf, T.: Diffusers: state-of-the-art diffusion models. https:\/\/github.com\/huggingface\/diffusers (2022)"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04104-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04104-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04104-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,24]],"date-time":"2025-09-24T14:00:17Z","timestamp":1758722417000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04104-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,23]]},"references-count":70,"journal-issue":{"issue":"13","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["4104"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04104-3","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2025,7,23]]},"assertion":[{"value":"2 July 2025","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 July 2025","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}