{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:08:55Z","timestamp":1750219735039,"version":"3.41.0"},"reference-count":68,"publisher":"Association for Computing Machinery (ACM)","issue":"3","license":[{"start":{"date-parts":[[2023,10,23]],"date-time":"2023-10-23T00:00:00Z","timestamp":1698019200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Multimedia Comput. Commun. Appl."],"published-print":{"date-parts":[[2024,3,31]]},"abstract":"<jats:p>VTON (Virtual Try-ON), as an innovative visual application in e-commerce scenarios with great commercial value, has been widely studied in recent years. Due to its better robustness and realistic effect, deformation-synthesize-based VTON has become the dominant approach in this field. Existing clothing deformation techniques optimize the mapping relations between the original clothing image and the ground truth (GT) image of the worn clothing. However, there are color differences between the original and GT clothing images caused by lighting, warping, and occlusion. The color differences may lead to misaligned clothing mapping by only minimizing the cost of pixel value difference. Another drawback is that taking the parsing prediction as GT will bring alignment remnant, rooting in the processing order of parsing and deformation. Aiming above two drawbacks, we put forward SAME-VTON (Self-Adaptive clothing Mapping basEd Virtual Try-ON) for achieving realistic virtual try-on results. The core of SAME-VTON is the self-adaptive clothing mapping technique, composed of two parts: a color-adaptive clothing mapping module and a parsing-adaptive prediction process. In the color-adaptive clothing mapping module, we map each pixel of the target clothing with a combination of multiple pixel values from the original clothing image, which considers both the position and color changes. Furthermore, different combination weights are learned to increase the diversity of color mapping. In the parsing-adaptive prediction process, the color-adaptive clothing mapping module is adopted to deform clothing first, then the human parsing result is predicted under the reference of the deformed clothing, which can avoid alignment remnant. Extensive experiments demonstrate that the proposed SAME-VTON with the self-adaptive clothing mapping technique can achieve optimal mapping in the case of large color differences and obtain superior results compared with existing deformation-synthesize-based VTON.<\/jats:p>","DOI":"10.1145\/3613453","type":"journal-article","created":{"date-parts":[[2023,8,8]],"date-time":"2023-08-08T12:02:47Z","timestamp":1691496167000},"page":"1-26","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":3,"title":["Self-Adaptive Clothing Mapping Based Virtual Try-on"],"prefix":"10.1145","volume":"20","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7223-8921","authenticated-orcid":false,"given":"Chengji","family":"Shen","sequence":"first","affiliation":[{"name":"Zhejiang University, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-3846-4977","authenticated-orcid":false,"given":"Zhenjiang","family":"Liu","sequence":"additional","affiliation":[{"name":"Alibaba Group, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-0543-0635","authenticated-orcid":false,"given":"Xin","family":"Gao","sequence":"additional","affiliation":[{"name":"Alibaba Group, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8640-8434","authenticated-orcid":false,"given":"Zunlei","family":"Feng","sequence":"additional","affiliation":[{"name":"Zhejiang University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2621-6048","authenticated-orcid":false,"given":"Mingli","family":"Song","sequence":"additional","affiliation":[{"name":"Zhejiang University, China"}]}],"member":"320","published-online":{"date-parts":[[2023,10,23]]},"reference":[{"key":"e_1_3_3_2_2","doi-asserted-by":"publisher","DOI":"10.1145\/3447239"},{"key":"e_1_3_3_3_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01391"},{"key":"e_1_3_3_4_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00538"},{"key":"e_1_3_3_5_2","first-page":"14638","volume-title":"Proceedings of the IEEE International Conference on Computer Vision","author":"Cui Aiyu","year":"2021","unstructured":"Aiyu Cui, Daniel McKee, and Svetlana Lazebnik. 2021. Dressing in order: Recurrent person image generation for pose transfer, virtual try-on and outfit editing. In Proceedings of the IEEE International Conference on Computer Vision. 14638\u201314647."},{"key":"e_1_3_3_6_2","doi-asserted-by":"publisher","DOI":"10.1145\/3531017"},{"key":"e_1_3_3_7_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"e_1_3_3_8_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.isprsjprs.2020.01.013"},{"key":"e_1_3_3_9_2","doi-asserted-by":"publisher","DOI":"10.1109\/WACV51458.2022.00226"},{"key":"e_1_3_3_10_2","doi-asserted-by":"publisher","DOI":"10.1145\/3326332"},{"key":"e_1_3_3_11_2","doi-asserted-by":"publisher","DOI":"10.1145\/3491226"},{"key":"e_1_3_3_12_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR48806.2021.9412052"},{"key":"e_1_3_3_13_2","doi-asserted-by":"publisher","DOI":"10.1145\/3474085.3475210"},{"key":"e_1_3_3_14_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01665"},{"key":"e_1_3_3_15_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00838"},{"key":"e_1_3_3_16_2","article-title":"Generative adversarial nets","volume":"27","author":"Goodfellow Ian","year":"2014","unstructured":"Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative adversarial nets. In Advances in Neural Information Processing Systems 27 (NIPS \u201914).","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_3_17_2","doi-asserted-by":"publisher","DOI":"10.1145\/3545610"},{"key":"e_1_3_3_18_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.01057"},{"key":"e_1_3_3_19_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00787"},{"key":"e_1_3_3_20_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_3_21_2","first-page":"512","volume-title":"Proceedings of the European Conference on Computer Vision","author":"Heming Zhu","year":"2020","unstructured":"Zhu Heming, Cao Yu, Jin Hang, Chen Weikai, Du Dong, Wang Zhangye, Cui Shuguang, and Han Xiaoguang. 2020. Deep Fashion3D: A dataset and benchmark for 3D garment reconstruction from single images. In Proceedings of the European Conference on Computer Vision. 512\u2013530."},{"key":"e_1_3_3_22_2","doi-asserted-by":"publisher","DOI":"10.5555\/3295222.3295408"},{"key":"e_1_3_3_23_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR.2010.579"},{"key":"e_1_3_3_24_2","doi-asserted-by":"publisher","DOI":"10.1145\/3343031.3351075"},{"key":"e_1_3_3_25_2","article-title":"Image-to-image translation with conditional adversarial networks","author":"Isola Phillip","year":"2017","unstructured":"Phillip Isola, Jun-Yan Zhu, Tinghui \u201cZhou\u201d, and Alexei A. Efros. 2017. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.","journal-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition."},{"key":"e_1_3_3_26_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58565-5_37"},{"key":"e_1_3_3_27_2","volume-title":"Proceedings of the International Conference on Learning Representations","author":"Jang Eric","year":"2017","unstructured":"Eric Jang, Shixiang Gu, and Ben Poole. 2017. Categorical reparameterization with Gumbel-softmax. In Proceedings of the International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=rkE3y85ee"},{"key":"e_1_3_3_28_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58565-5_2"},{"key":"e_1_3_3_29_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01545"},{"key":"e_1_3_3_30_2","volume-title":"Proceedings of the International Conference on Learning Representations","author":"Kingma Diederik P.","year":"2015","unstructured":"Diederik P. Kingma and Jimmy Ba. 2015. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations. http:\/\/arxiv.org\/abs\/1412.6980"},{"key":"e_1_3_3_31_2","doi-asserted-by":"publisher","DOI":"10.1145\/3450626.3459884"},{"key":"e_1_3_3_32_2","article-title":"VOGUE: Try-on by StyleGAN interpolation optimization","author":"Lewis Kathleen M.","year":"2021","unstructured":"Kathleen M. Lewis, Srivatsan Varadharajan, and Ira Kemelmacher-Shlizerman. 2021. VOGUE: Try-on by StyleGAN interpolation optimization. arXiv preprint arXiv:2101.02285 (2021).","journal-title":"arXiv preprint arXiv:2101.02285"},{"key":"e_1_3_3_33_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01529"},{"key":"e_1_3_3_34_2","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2020.3048039"},{"key":"e_1_3_3_35_2","doi-asserted-by":"publisher","DOI":"10.1145\/2659234"},{"key":"e_1_3_3_36_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00366"},{"key":"e_1_3_3_37_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00650"},{"key":"e_1_3_3_38_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00513"},{"key":"e_1_3_3_39_2","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshop","author":"Minar Matiur Rahman","year":"2020","unstructured":"Matiur Rahman Minar, Thai Thanh Tuan, Heejune Ahn, Paul Rosin, and Yu-Kun Lai. 2020. 3D reconstruction of clothes using a human body model and its application to image-based virtual try-on. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshop."},{"key":"e_1_3_3_40_2","first-page":"4","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshop","author":"Minar Matiur Rahman","year":"2020","unstructured":"Matiur Rahman Minar, Thai Thanh Tuan, Heejune Ahn, Paul L. Rosin, and Yu-Kun Lai. 2020. CP-VTON+: Clothing shape and texture preserving image-based virtual try-on. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshop. 4."},{"key":"e_1_3_3_41_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00705"},{"key":"e_1_3_3_42_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW56347.2022.00243"},{"key":"e_1_3_3_43_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01112"},{"key":"e_1_3_3_44_2","first-page":"8024","volume-title":"Advances in Neural Information Processing Systems 32 (NIPS\u201919)","author":"Paszke Adam","year":"2019","unstructured":"Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019. PyTorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems 32 (NIPS\u201919). 8024\u20138035. http:\/\/papers.neurips.cc\/paper\/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf"},{"key":"e_1_3_3_45_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00739"},{"key":"e_1_3_3_46_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00771"},{"key":"e_1_3_3_47_2","article-title":"DensePose: Dense human pose estimation in the wild","author":"Neverova Iasonas Kokkinos R{i}za Alp G\u00fcler, Natalia","year":"2018","unstructured":"Iasonas Kokkinos R{i}za Alp G\u00fcler, Natalia Neverova. 2018. DensePose: Dense human pose estimation in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.","journal-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition"},{"key":"e_1_3_3_48_2","doi-asserted-by":"publisher","DOI":"10.1109\/cvpr.2017.12"},{"key":"e_1_3_3_49_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"e_1_3_3_50_2","unstructured":"Tim Salimans Ian Goodfellow Wojciech Zaremba Vicki Cheung Alec Radford and Xi Chen. 2016. Improved techniques for training GANs. arXiv:cs.LG\/1606.03498 (2016)."},{"key":"e_1_3_3_51_2","article-title":"Self-supervised collision handling via generative 3D garment models for virtual try-on","author":"Santesteban Igor","year":"2021","unstructured":"Igor Santesteban, Nils Thuerey, Miguel A. Otaduy, and Dan Casas. 2021. Self-supervised collision handling via generative 3D garment models for virtual try-on. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.","journal-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition."},{"key":"e_1_3_3_52_2","doi-asserted-by":"crossref","unstructured":"Scott Schaefer Travis McPhail and Joe Warren. 2006. Image deformation using moving least squares. ACM Trans. Graph. 25 3 (2006) 533\u2013540.","DOI":"10.1145\/1141911.1141920"},{"key":"e_1_3_3_53_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58523-5_14"},{"key":"e_1_3_3_54_2","volume-title":"Proceedings of the International Conference on Learning Representations","author":"Simonyan K.","year":"2015","unstructured":"K. Simonyan and A. Zisserman. 2015. Very deep convolutional networks for large-scale image recognition. In Proceedings of the International Conference on Learning Representations."},{"key":"e_1_3_3_55_2","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2022.3170728"},{"key":"e_1_3_3_56_2","first-page":"109","volume-title":"Proceedings of the Symposium on Geometry Processing","volume":"4","author":"Sorkine Olga","year":"2007","unstructured":"Olga Sorkine and Marc Alexa. 2007. As-rigid-as-possible surface modeling. In Proceedings of the Symposium on Geometry Processing, Vol. 4. 109\u2013116."},{"key":"e_1_3_3_57_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58580-8_1"},{"key":"e_1_3_3_58_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01261-8_36"},{"key":"e_1_3_3_59_2","volume-title":"arXiv:1912.06324 [Cs]","author":"Wang Jiahang","year":"2020","unstructured":"Jiahang Wang, Wei Zhang, Weizhong Liu, and Tao Mei. 2020. Down to the last detail: Virtual try-on with detail carving. arXiv:1912.06324 [Cs] (2020)."},{"key":"e_1_3_3_60_2","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2003.819861"},{"key":"e_1_3_3_61_2","doi-asserted-by":"publisher","DOI":"10.1111\/1467-9868.00374"},{"key":"e_1_3_3_62_2","volume-title":"Advances in Neural Information Processing Systems (NIPS\u201921)","author":"Xie Zhenyu","year":"2021","unstructured":"Zhenyu Xie, Zaiyu Huang, Fuwei Zhao, Haoye Dong, Michael Kampffmeyer, and Xiaodan Liang. 2021. Towards scalable unpaired virtual try-on via patch-routed spatially-adaptive GAN. In Advances in Neural Information Processing Systems (NIPS\u201921)."},{"key":"e_1_3_3_63_2","article-title":"WAS-VTON: Warping architecture search for virtual try-on network","author":"Xie Zhenyu","year":"2021","unstructured":"Zhenyu Xie, Xujie Zhang, Fuwei Zhao, Haoye Dong, Michael C. Kampffmeyer, Haonan Yan, and Xiaodan Liang. 2021. WAS-VTON: Warping architecture search for virtual try-on network. In Proceedings of the ACM International Conference on Multimedia (MM\u201921). 3350\u20133359.","journal-title":"Proceedings of the ACM International Conference on Multimedia (MM\u201921)."},{"key":"e_1_3_3_64_2","doi-asserted-by":"publisher","DOI":"10.1145\/3554739"},{"key":"e_1_3_3_65_2","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","author":"Yang Han","year":"2020","unstructured":"Han Yang, Ruimao Zhang, Xiaobao Guo, Wei Liu, Wangmeng Zuo, and Ping Luo. 2020. Towards photo-realistic virtual try-on by adaptively generating \\(\\leftrightarrow\\) preserving image content. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition."},{"key":"e_1_3_3_66_2","doi-asserted-by":"publisher","DOI":"10.1145\/3425636"},{"key":"e_1_3_3_67_2","first-page":"25739","article-title":"Deep model reassembly","author":"Yang Xingyi","year":"2022","unstructured":"Xingyi Yang, Daquan Zhou, Songhua Liu, Jingwen Ye, and Xinchao Wang. 2022. Deep model reassembly. In Advances in Neural Information Processing Systems 35 (NIPS\u201922).25739\u201325753.","journal-title":"Advances in Neural Information Processing Systems 35 (NIPS\u201922)."},{"key":"e_1_3_3_68_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.01061"},{"key":"e_1_3_3_69_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01299"}],"container-title":["ACM Transactions on Multimedia Computing, Communications, and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3613453","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3613453","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:36:30Z","timestamp":1750178190000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3613453"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,23]]},"references-count":68,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2024,3,31]]}},"alternative-id":["10.1145\/3613453"],"URL":"https:\/\/doi.org\/10.1145\/3613453","relation":{},"ISSN":["1551-6857","1551-6865"],"issn-type":[{"type":"print","value":"1551-6857"},{"type":"electronic","value":"1551-6865"}],"subject":[],"published":{"date-parts":[[2023,10,23]]},"assertion":[{"value":"2022-10-12","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2023-07-27","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2023-10-23","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}