{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T22:10:52Z","timestamp":1776204652586,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":37,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,12,10]],"date-time":"2023-12-10T00:00:00Z","timestamp":1702166400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Hong Kong Research Grants Council (RGC) GRF Scheme","award":["CityU 11216122"],"award-info":[{"award-number":["CityU 11216122"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,12,10]]},"DOI":"10.1145\/3610548.3618232","type":"proceedings-article","created":{"date-parts":[[2023,12,11]],"date-time":"2023-12-11T12:28:40Z","timestamp":1702297720000},"page":"1-11","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":6,"title":["Text-Guided Vector Graphics Customization"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3517-808X","authenticated-orcid":false,"given":"Peiying","family":"Zhang","sequence":"first","affiliation":[{"name":"City University of Hong Kong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4007-2776","authenticated-orcid":false,"given":"Nanxuan","family":"Zhao","sequence":"additional","affiliation":[{"name":"Adobe Research, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7014-5377","authenticated-orcid":false,"given":"Jing","family":"Liao","sequence":"additional","affiliation":[{"name":"City University of Hong Kong, China"}]}],"member":"320","published-online":{"date-parts":[[2023,12,11]]},"reference":[{"key":"e_1_3_2_2_1_1","volume-title":"The fundamentals of graphic design","author":"Ambrose Gavin","unstructured":"Gavin Ambrose, Paul Harris, and Nigel Ball. 2019. The fundamentals of graphic design. Bloomsbury Publishing."},{"key":"e_1_3_2_2_2_1","volume-title":"Deep vit features as dense visual descriptors. arXiv preprint arXiv:2112.05814 2, 3","author":"Amir Shir","year":"2021","unstructured":"Shir Amir, Yossi Gandelsman, Shai Bagon, and Tali Dekel. 2021. Deep vit features as dense visual descriptors. arXiv preprint arXiv:2112.05814 2, 3 (2021), 4."},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00951"},{"key":"e_1_3_2_2_4_1","volume-title":"What is in a Text-to-Image Prompt: The Potential of Stable Diffusion in Visual Arts Education. arXiv preprint arXiv:2301.01902","author":"Dehouche Nassim","year":"2023","unstructured":"Nassim Dehouche and Kullathida. 2023. What is in a Text-to-Image Prompt: The Potential of Stable Diffusion in Visual Arts Education. arXiv preprint arXiv:2301.01902 (2023)."},{"key":"e_1_3_2_2_5_1","volume-title":"Bayesian Image Vectorization: the probabilistic inversion of vector image rasterization","author":"Diebel James\u00a0Richard","unstructured":"James\u00a0Richard Diebel. 2008. Bayesian Image Vectorization: the probabilistic inversion of vector image rasterization. Stanford University."},{"key":"e_1_3_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3386569.3392401"},{"key":"e_1_3_2_2_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3130800.3130888"},{"key":"e_1_3_2_2_8_1","first-page":"5207","article-title":"Clipdraw: Exploring text-to-drawing synthesis through language-image encoders","volume":"35","author":"Frans Kevin","year":"2022","unstructured":"Kevin Frans, Lisa Soros, and Olaf Witkowski. 2022. Clipdraw: Exploring text-to-drawing synthesis through language-image encoders. Advances in Neural Information Processing Systems 35 (2022), 5207\u20135218.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_9_1","volume-title":"An image is worth one word: Personalizing text-to-image generation using textual inversion. arXiv preprint arXiv:2208.01618","author":"Gal Rinon","year":"2022","unstructured":"Rinon Gal, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit\u00a0H Bermano, Gal Chechik, and Daniel Cohen-Or. 2022. An image is worth one word: Personalizing text-to-image generation using textual inversion. arXiv preprint arXiv:2208.01618 (2022)."},{"key":"e_1_3_2_2_10_1","volume-title":"A neural representation of sketch drawings. arXiv preprint arXiv:1704.03477","author":"Ha David","year":"2017","unstructured":"David Ha and Douglas Eck. 2017. A neural representation of sketch drawings. arXiv preprint arXiv:1704.03477 (2017)."},{"key":"e_1_3_2_2_11_1","volume-title":"Conference on Robot Learning. PMLR, 550\u2013560","author":"Hadjivelichkov Denis","year":"2023","unstructured":"Denis Hadjivelichkov, Sicelukwanda Zwane, Lourdes Agapito, Marc\u00a0Peter Deisenroth, and Dimitrios Kanoulas. 2023. One-Shot Transfer of Affordance Regions? AffCorrs!. In Conference on Robot Learning. PMLR, 550\u2013560."},{"key":"e_1_3_2_2_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3197517.3201312"},{"key":"e_1_3_2_2_13_1","volume-title":"VectorFusion: Text-to-SVG by Abstracting Pixel-Based Diffusion Models. arXiv preprint arXiv:2211.11319","author":"Jain Ajay","year":"2022","unstructured":"Ajay Jain, Amber Xie, and Pieter Abbeel. 2022. VectorFusion: Text-to-SVG by Abstracting Pixel-Based Diffusion Models. arXiv preprint arXiv:2211.11319 (2022)."},{"key":"e_1_3_2_2_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/1964921.1964994"},{"key":"e_1_3_2_2_15_1","volume-title":"Multi-Concept Customization of Text-to-Image Diffusion. arXiv preprint arXiv:2212.04488","author":"Kumari Nupur","year":"2022","unstructured":"Nupur Kumari, Bingliang Zhang, Richard Zhang, Eli Shechtman, and Jun-Yan Zhu. 2022. Multi-Concept Customization of Text-to-Image Diffusion. arXiv preprint arXiv:2212.04488 (2022)."},{"key":"e_1_3_2_2_16_1","volume-title":"Controllable text-to-image generation. Advances in Neural Information Processing Systems 32","author":"Li Bowen","year":"2019","unstructured":"Bowen Li, Xiaojuan Qi, Thomas Lukasiewicz, and Philip Torr. 2019. Controllable text-to-image generation. Advances in Neural Information Processing Systems 32 (2019)."},{"key":"e_1_3_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3414685.3417763"},{"key":"e_1_3_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01583"},{"key":"e_1_3_2_2_19_1","volume-title":"Point set registration: Coherent point drift","author":"Myronenko Andriy","year":"2010","unstructured":"Andriy Myronenko and Xubo Song. 2010. Point set registration: Coherent point drift. IEEE transactions on pattern analysis and machine intelligence 32, 12 (2010), 2262\u20132275."},{"key":"e_1_3_2_2_20_1","volume-title":"Glide: Towards photorealistic image generation and editing with text-guided diffusion models. arXiv preprint arXiv:2112.10741","author":"Nichol Alex","year":"2021","unstructured":"Alex Nichol, Prafulla Dhariwal, Aditya Ramesh, Pranav Shyam, Pamela Mishkin, Bob McGrew, Ilya Sutskever, and Mark Chen. 2021. Glide: Towards photorealistic image generation and editing with text-guided diffusion models. arXiv preprint arXiv:2112.10741 (2021)."},{"key":"e_1_3_2_2_21_1","volume-title":"U2-Net: Going deeper with nested U-structure for salient object detection. Pattern recognition 106","author":"Qin Xuebin","year":"2020","unstructured":"Xuebin Qin, Zichen Zhang, Chenyang Huang, Masood Dehghan, Osmar\u00a0R Zaiane, and Martin Jagersand. 2020. U2-Net: Going deeper with nested U-structure for salient object detection. Pattern recognition 106 (2020), 107404."},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/MMUL.2003.1218261"},{"key":"e_1_3_2_2_23_1","volume-title":"International conference on machine learning. PMLR, 8748\u20138763","author":"Radford Alec","year":"2021","unstructured":"Alec Radford, Jong\u00a0Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, 2021. Learning transferable visual models from natural language supervision. In International conference on machine learning. PMLR, 8748\u20138763."},{"key":"e_1_3_2_2_24_1","volume-title":"International Conference on Machine Learning. PMLR, 8821\u20138831","author":"Ramesh Aditya","year":"2021","unstructured":"Aditya Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Radford, Mark Chen, and Ilya Sutskever. 2021. Zero-shot text-to-image generation. In International Conference on Machine Learning. PMLR, 8821\u20138831."},{"key":"e_1_3_2_2_25_1","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 7342\u20137351","author":"Reddy Pradyumna","year":"2021","unstructured":"Pradyumna Reddy, Michael Gharbi, Michal Lukac, and Niloy\u00a0J Mitra. 2021. Im2vec: Synthesizing vector graphics without vector supervision. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 7342\u20137351."},{"key":"e_1_3_2_2_26_1","volume-title":"International conference on machine learning. PMLR, 1060\u20131069","author":"Reed Scott","year":"2016","unstructured":"Scott Reed, Zeynep Akata, Xinchen Yan, Lajanugen Logeswaran, Bernt Schiele, and Honglak Lee. 2016. Generative adversarial text to image synthesis. In International conference on machine learning. PMLR, 1060\u20131069."},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01416"},{"key":"e_1_3_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"e_1_3_2_2_29_1","volume-title":"Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation. arXiv preprint arXiv:2208.12242","author":"Ruiz Nataniel","year":"2022","unstructured":"Nataniel Ruiz, Yuanzhen Li, Varun Jampani, Yael Pritch, Michael Rubinstein, and Kfir Aberman. 2022. Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation. arXiv preprint arXiv:2208.12242 (2022)."},{"key":"e_1_3_2_2_30_1","volume-title":"Styleclipdraw: Coupling content and style in text-to-drawing translation. arXiv preprint arXiv:2202.12362","author":"Schaldenbrand Peter","year":"2022","unstructured":"Peter Schaldenbrand, Zhixuan Liu, and Jean Oh. 2022. Styleclipdraw: Coupling content and style in text-to-drawing translation. arXiv preprint arXiv:2202.12362 (2022)."},{"key":"e_1_3_2_2_31_1","unstructured":"Peter Selinger. 2003. Potrace: a polygon-based tracing algorithm."},{"key":"e_1_3_2_2_32_1","volume-title":"CLIPVG: Text-Guided Image Manipulation Using Differentiable Vector Graphics. arXiv preprint arXiv:2212.02122","author":"Song Yiren","year":"2022","unstructured":"Yiren Song, Xning Shao, Kang Chen, Weidong Zhang, Minzhe Li, and Zhongliang Jing. 2022. CLIPVG: Text-Guided Image Manipulation Using Differentiable Vector Graphics. arXiv preprint arXiv:2212.02122 (2022)."},{"key":"e_1_3_2_2_33_1","volume-title":"Art and Design: 11th International Conference, EvoMUSART 2022, Held as Part of EvoStar 2022, Madrid, Spain, April 20\u201322, 2022, Proceedings. Springer, 275\u2013291","author":"Tian Yingtao","year":"2022","unstructured":"Yingtao Tian and David Ha. 2022. Modern evolution strategies for creativity: Fitting concrete images and abstract concepts. In Artificial Intelligence in Music, Sound, Art and Design: 11th International Conference, EvoMUSART 2022, Held as Part of EvoStar 2022, Madrid, Spain, April 20\u201322, 2022, Proceedings. Springer, 275\u2013291."},{"key":"e_1_3_2_2_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/3528223.3530068"},{"key":"e_1_3_2_2_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/1390156.1390297"},{"key":"e_1_3_2_2_36_1","volume-title":"Computer Graphics Forum, Vol.\u00a042","author":"Yang Jinfan","unstructured":"Jinfan Yang, Nicholas Vining, Shakiba Kheradmand, Nathan Carr, Leonid Sigal, and Alla Sheffer. 2023. Subpixel Deblurring of Anti-Aliased Raster Clip-Art. In Computer Graphics Forum, Vol.\u00a042. Wiley Online Library, 61\u201376."},{"key":"e_1_3_2_2_37_1","volume-title":"Effective clipart image vectorization through direct optimization of bezigons","author":"Yang Ming","year":"2015","unstructured":"Ming Yang, Hongyang Chao, Chi Zhang, Jun Guo, Lu Yuan, and Jian Sun. 2015. Effective clipart image vectorization through direct optimization of bezigons. IEEE transactions on visualization and computer graphics 22, 2 (2015), 1063\u20131075."}],"event":{"name":"SA '23: SIGGRAPH Asia 2023","location":"Sydney NSW Australia","acronym":"SA '23","sponsor":["SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques"]},"container-title":["SIGGRAPH Asia 2023 Conference Papers"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3610548.3618232","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3610548.3618232","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T09:29:58Z","timestamp":1755768598000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3610548.3618232"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,10]]},"references-count":37,"alternative-id":["10.1145\/3610548.3618232","10.1145\/3610548"],"URL":"https:\/\/doi.org\/10.1145\/3610548.3618232","relation":{},"subject":[],"published":{"date-parts":[[2023,12,10]]},"assertion":[{"value":"2023-12-11","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}