{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,25]],"date-time":"2026-03-25T14:47:14Z","timestamp":1774450034552,"version":"3.50.1"},"reference-count":38,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,12,14]],"date-time":"2025-12-14T00:00:00Z","timestamp":1765670400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,12,14]],"date-time":"2025-12-14T00:00:00Z","timestamp":1765670400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"The Open Research Fund of Hubei Provincial Engineering Research Center for Intelligent Textile and Fashion, Wuhan Textile University","award":["2023HBITF02"],"award-info":[{"award-number":["2023HBITF02"]}]},{"name":"The Open Research Fund of Hubei Provincial Engineering Research Center for Intelligent Textile and Fashion, Wuhan Textile University","award":["2023HBITF02"],"award-info":[{"award-number":["2023HBITF02"]}]},{"name":"The Open Research Fund of Hubei Provincial Engineering Research Center for Intelligent Textile and Fashion, Wuhan Textile University","award":["2023HBITF02"],"award-info":[{"award-number":["2023HBITF02"]}]},{"name":"The Open Research Fund of Hubei Provincial Engineering Research Center for Intelligent Textile and Fashion, Wuhan Textile University","award":["2023HBITF02"],"award-info":[{"award-number":["2023HBITF02"]}]},{"name":"The Open Research Fund of Hubei Provincial Engineering Research Center for Intelligent Textile and Fashion, Wuhan Textile University","award":["2023HBITF02"],"award-info":[{"award-number":["2023HBITF02"]}]},{"name":"The Open Research Fund of Hubei Provincial Engineering Research Center for Intelligent Textile and Fashion, Wuhan Textile University","award":["2023HBITF02"],"award-info":[{"award-number":["2023HBITF02"]}]},{"name":"The Open Research Fund of Hubei Provincial Engineering Research Center for Intelligent Textile and Fashion, Wuhan Textile University","award":["2023HBITF02"],"award-info":[{"award-number":["2023HBITF02"]}]},{"name":"The Open Research Fund of Hubei Provincial Engineering Research Center for Intelligent Textile and Fashion, Wuhan Textile University","award":["2023HBITF02"],"award-info":[{"award-number":["2023HBITF02"]}]},{"name":"The Open Research Fund of Hubei Provincial Engineering Research Center for Intelligent Textile and Fashion, Wuhan Textile University","award":["2023HBITF02"],"award-info":[{"award-number":["2023HBITF02"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2026,1]]},"DOI":"10.1007\/s00371-025-04223-x","type":"journal-article","created":{"date-parts":[[2025,12,14]],"date-time":"2025-12-14T03:42:52Z","timestamp":1765683772000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Har-vton: a diffusion-based virtual try-on framework with hybrid attention and receptive field modules"],"prefix":"10.1007","volume":"42","author":[{"given":"Yulin","family":"Xiong","sequence":"first","affiliation":[]},{"given":"Yuxin","family":"Hong","sequence":"additional","affiliation":[]},{"given":"Xuyan","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Jianlin","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Zimao","family":"Li","sequence":"additional","affiliation":[]},{"given":"Ruhan","family":"He","sequence":"additional","affiliation":[]},{"given":"Xiao","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Bo","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Meng","family":"Shi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,12,14]]},"reference":[{"key":"4223_CR1","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11263-025-02570-9","volume":"133","author":"X Li","year":"2025","unstructured":"Li, X., Ren, Y., Jin, X., Lan, C., Wang, X., Zeng, W., Wang, X., Chen, Z.: Diffusion models for image restoration and enhancement: a comprehensive survey. Int. J. Comput. Vis. 133, 1\u201331 (2025)","journal-title":"Int. J. Comput. Vis."},{"key":"4223_CR2","doi-asserted-by":"crossref","unstructured":"Huang, Y., Huang, J., Liu, Y., Yan, M., Lv, J., Liu, J., Xiong, W., Zhang, H., Cao, L., Chen, S.: Diffusion model-based image editing: a survey. IEEE Transactions on Pattern Analysis and Machine Intelligence (2025)","DOI":"10.1109\/TPAMI.2025.3541625"},{"key":"4223_CR3","doi-asserted-by":"crossref","unstructured":"Goel, P., Sharma, B., Kumari, A., Gupta, A.K., Chhajed, P.: A review on virtual try-on. In: 2024 15th International Conference on Computing Communication and Networking Technologies (ICCCNT), pp. 1\u20135 (2024)","DOI":"10.1109\/ICCCNT61001.2024.10724088"},{"key":"4223_CR4","doi-asserted-by":"crossref","unstructured":"Choi, S., Park, S., Lee, M., Choo, J.: Viton-hd: High-resolution virtual try-on via misalignment-aware normalization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14131\u201314140 (2021)","DOI":"10.1109\/CVPR46437.2021.01391"},{"key":"4223_CR5","doi-asserted-by":"crossref","unstructured":"Jetchev, N., Bergmann, U.: The conditional analogy gan: swapping fashion articles on people images. In: Proceedings of the IEEE International Conference on Computer Vision Workshops, pp. 2287\u20132292 (2017)","DOI":"10.1109\/ICCVW.2017.269"},{"key":"4223_CR6","doi-asserted-by":"crossref","unstructured":"Xie, Z., Huang, Z., Dong, X., Zhao, F., Dong, H., Zhang, X., Zhu, F., Liang, X.: Gp-vton: Towards general purpose virtual try-on via collaborative local-flow global-parsing learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 23550\u201323559 (2023)","DOI":"10.1109\/CVPR52729.2023.02255"},{"key":"4223_CR7","doi-asserted-by":"crossref","unstructured":"Han, X., Wu, Z., Wu, Z., Yu, R., Davis, L.S.: Viton: An image-based virtual try-on network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7543\u20137552 (2018)","DOI":"10.1109\/CVPR.2018.00787"},{"key":"4223_CR8","doi-asserted-by":"crossref","unstructured":"Wang, B., Zheng, H., Liang, X., Chen, Y., Lin, L., Yang, M.: Toward characteristic-preserving image-based virtual try-on network. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 589\u2013604 (2018)","DOI":"10.1007\/978-3-030-01261-8_36"},{"key":"4223_CR9","doi-asserted-by":"crossref","unstructured":"Lee, S., Gu, G., Park, S., Choi, S., Choo, J.: High-resolution virtual try-on with misalignment and occlusion-handled conditions. In: European Conference on Computer Vision, pp. 204\u2013219 (2022)","DOI":"10.1007\/978-3-031-19790-1_13"},{"key":"4223_CR10","doi-asserted-by":"crossref","unstructured":"Han, X., Hu, X., Huang, W., Scott, M.R.: Clothflow: A flow-based model for clothed person generation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10471\u201310480 (2019)","DOI":"10.1109\/ICCV.2019.01057"},{"key":"4223_CR11","doi-asserted-by":"crossref","unstructured":"Yang, H., Zhang, R., Guo, X., Liu, W., Zuo, W., Luo, P.: Towards photo-realistic virtual try-on by adaptively generating-preserving image content. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7850\u20137859 (2020)","DOI":"10.1109\/CVPR42600.2020.00787"},{"key":"4223_CR12","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10684\u201310695 (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"4223_CR13","doi-asserted-by":"crossref","unstructured":"Kim, J., Gu, G., Park, M., Park, S., Choo, J.: Stableviton: Learning semantic correspondence with latent diffusion model for virtual try-on. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8176\u20138185 (2024)","DOI":"10.1109\/CVPR52733.2024.00781"},{"key":"4223_CR14","doi-asserted-by":"crossref","unstructured":"Morelli, D., Baldrati, A., Cartella, G., Cornia, M., Bertini, M., Cucchiara, R.: Ladi-vton: Latent diffusion textual-inversion enhanced virtual try-on. In: Proceedings of the 31st ACM International Conference on Multimedia, pp. 8580\u20138589 (2023)","DOI":"10.1145\/3581783.3612137"},{"key":"4223_CR15","doi-asserted-by":"crossref","unstructured":"Zhu, L., Yang, D., Zhu, T., Reda, F., Chan, W., Saharia, C., Norouzi, M., Kemelmacher-Shlizerman, I.: Tryondiffusion: a tale of two unets. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4606\u20134615 (2023)","DOI":"10.1109\/CVPR52729.2023.00447"},{"key":"4223_CR16","doi-asserted-by":"crossref","unstructured":"Xu, Y., Gu, T., Chen, W., Chen, C.: Ootdiffusion: Outfitting fusion based latent diffusion for controllable virtual try-on. Preprint at arXiv:2403.01779 (2024)","DOI":"10.1609\/aaai.v39i9.32973"},{"key":"4223_CR17","doi-asserted-by":"crossref","unstructured":"Choi, Y., Kwak, S., Lee, K., Choi, H., Shin, J.: Improving diffusion models for authentic virtual try-on in the wild. In: European Conference on Computer Vision, pp. 206\u2013235 (2024)","DOI":"10.1007\/978-3-031-73016-0_13"},{"key":"4223_CR18","doi-asserted-by":"crossref","unstructured":"Shen, F., Jiang, X., He, X., Ye, H., Wang, C., Du, X., Li, Z., Tang, J.: Imagdressing-v1: Customizable virtual dressing. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 39, pp. 6795\u20136804 (2025)","DOI":"10.1609\/aaai.v39i7.32729"},{"key":"4223_CR19","first-page":"6840","volume":"33","author":"J Ho","year":"2020","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. Adv. Neural. Inf. Process. Syst. 33, 6840\u20136851 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"4223_CR20","doi-asserted-by":"crossref","unstructured":"Mou, C., Wang, X., Xie, L., Wu, Y., Zhang, J., Qi, Z., Shan, Y.: T2i-adapter: Learning adapters to dig out more controllable ability for text-to-image diffusion models. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, pp. 4296\u20134304 (2024)","DOI":"10.1609\/aaai.v38i5.28226"},{"key":"4223_CR21","doi-asserted-by":"crossref","unstructured":"Zhang, L., Rao, A., Agrawala, M.: Adding conditional control to text-to-image diffusion models. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3836\u20133847 (2023)","DOI":"10.1109\/ICCV51070.2023.00355"},{"key":"4223_CR22","unstructured":"Ye, H., Zhang, J., Liu, S., Han, X., Yang, W.: Ip-adapter: Text compatible image prompt adapter for text-to-image diffusion models. Preprint at arXiv:2308.06721 (2023)"},{"key":"4223_CR23","first-page":"11127","volume":"36","author":"S Zhao","year":"2023","unstructured":"Zhao, S., Chen, D., Chen, Y.-C., Bao, J., Hao, S., Yuan, L., Wong, K.-Y.K.: Uni-controlnet: all-in-one control to text-to-image diffusion models. Adv. Neural. Inf. Process. Syst. 36, 11127\u201311150 (2023)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"3","key":"4223_CR24","doi-asserted-by":"publisher","first-page":"464","DOI":"10.1038\/s41433-023-02724-4","volume":"38","author":"Y Huang","year":"2024","unstructured":"Huang, Y., Cheung, C.Y., Li, D., Tham, Y.C., Sheng, B., Cheng, C.Y., Wang, Y.X., Wong, T.Y.: AI-integrated ocular imaging for predicting cardiovascular disease: advancements and future outlook. Eye 38(3), 464\u2013472 (2024)","journal-title":"Eye"},{"issue":"3\u20134","key":"4223_CR25","doi-asserted-by":"publisher","first-page":"2011","DOI":"10.1002\/cav.2011","volume":"32","author":"Y Zhou","year":"2021","unstructured":"Zhou, Y., Chen, Z., Sheng, B., Li, P., Kim, J., Wu, E.: Aff-dehazing: attention-based feature fusion network for low-light image dehazing. Comput. Anim. Virtual Worlds 32(3\u20134), 2011 (2021)","journal-title":"Comput. Anim. Virtual Worlds"},{"issue":"6","key":"4223_CR26","doi-asserted-by":"publisher","first-page":"538","DOI":"10.1016\/j.vrih.2023.06.008","volume":"5","author":"T Junjie","year":"2023","unstructured":"Junjie, T., Yinghui, W., Haomiao, M., Tao, Y., Lingyu, A., Zhang, S., Wei, L.: Image defocus deblurring method based on gradient difference of boundary neighborhood. Virtual Real. Intell. Hardw. 5(6), 538\u2013549 (2023)","journal-title":"Virtual Real. Intell. Hardw."},{"key":"4223_CR27","doi-asserted-by":"crossref","unstructured":"Cao, Z., Simon, T., Wei, S.-E., Sheikh, Y.: Realtime multi-person 2d pose estimation using part affinity fields. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7291\u20137299 (2017)","DOI":"10.1109\/CVPR.2017.143"},{"key":"4223_CR28","unstructured":"Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., et al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763 (2021)"},{"key":"4223_CR29","doi-asserted-by":"publisher","first-page":"3408","DOI":"10.1109\/TETCI.2024.3382233","volume":"8","author":"L Li","year":"2024","unstructured":"Li, L., Chen, Z., Dai, L., Li, R., Sheng, B.: MA-MFCNet: Mixed attention-based multi-scale feature calibration network for image dehazing. IEEE Trans. Emerg. Top. Comput. Intell. 8, 3408\u20133421 (2024)","journal-title":"IEEE Trans. Emerg. Top. Comput. Intell."},{"key":"4223_CR30","doi-asserted-by":"crossref","unstructured":"Chen, Z., Liu, X., Sheng, B., Li, P.: Garnet: graph attention residual networks based on adversarial learning for 3D human pose estimation. In: Computer Graphics International Conference, pp. 276\u2013287 (2020)","DOI":"10.1007\/978-3-030-61864-3_24"},{"key":"4223_CR31","doi-asserted-by":"crossref","unstructured":"Liu, R., Liu, M., Sheng, B., Li, H., Li, P., Song, H., Zhang, P., Jiang, L., Shen, D.: NHBS-net: A feature fusion attention network for ultrasound neonatal hip bone segmentation. IEEE Trans. Med. Imaging 40(12), 3446\u20133458 (2021)","DOI":"10.1109\/TMI.2021.3087857"},{"key":"4223_CR32","doi-asserted-by":"publisher","first-page":"1432","DOI":"10.1109\/TIP.2024.3364056","volume":"33","author":"Y Chen","year":"2024","unstructured":"Chen, Y., Jiang, R., Zheng, Y., Sheng, B., Yang, Z.-X., Wu, E.: Dual branch multi-level semantic learning for few-shot segmentation. IEEE Trans. Image Process. 33, 1432\u20131447 (2024)","journal-title":"IEEE Trans. Image Process."},{"key":"4223_CR33","doi-asserted-by":"crossref","unstructured":"Shen, B., Li, L., Hu, X., Guo, S., Huang, J., Liang, Z.: Point cloud upsampling generative adversarial network based on residual multi-scale off-set attention. Virtual Real. Intell. Hardw. 5(1), 81\u201391 (2023)","DOI":"10.1016\/j.vrih.2022.08.016"},{"key":"4223_CR34","doi-asserted-by":"crossref","unstructured":"Liu, S., Huang, D., et al.: Receptive field block net for accurate and fast object detection. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 385\u2013400 (2018)","DOI":"10.1007\/978-3-030-01252-6_24"},{"key":"4223_CR35","doi-asserted-by":"publisher","first-page":"81597","DOI":"10.1109\/ACCESS.2022.3195176","volume":"10","author":"Y Li","year":"2022","unstructured":"Li, Y., Fan, Y., Wang, S., Bai, J., Li, K.: Application of yolov5 based on attention mechanism and receptive field in identifying defects of thangka images. IEEE Access 10, 81597\u201381611 (2022)","journal-title":"IEEE Access"},{"issue":"4","key":"4223_CR36","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"key":"4223_CR37","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: Gans trained by a two time-scale update rule converge to a local nash equilibrium. In: Advances in Neural Information Processing Systems, vol. 30, pp. 6626\u20136637 (2017)"},{"key":"4223_CR38","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04223-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04223-x","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04223-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T13:03:23Z","timestamp":1772629403000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04223-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,14]]},"references-count":38,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,1]]}},"alternative-id":["4223"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04223-x","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,14]]},"assertion":[{"value":"5 June 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 November 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 December 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"62"}}