{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,20]],"date-time":"2026-03-20T16:01:19Z","timestamp":1774022479430,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":58,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,8,10]]},"DOI":"10.1145\/3721238.3730644","type":"proceedings-article","created":{"date-parts":[[2025,7,23]],"date-time":"2025-07-23T08:42:43Z","timestamp":1753260163000},"page":"1-9","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":3,"title":["Dynamic Concepts Personalization from Single Videos"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2177-8327","authenticated-orcid":false,"given":"Rameen","family":"Abdal","sequence":"first","affiliation":[{"name":"Snap Research, Palo Alto, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7757-6137","authenticated-orcid":false,"given":"Or","family":"Patashnik","sequence":"additional","affiliation":[{"name":"Snap Research, Palo Alto, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7611-9310","authenticated-orcid":false,"given":"Ivan","family":"Skorokhodov","sequence":"additional","affiliation":[{"name":"Snap Research, Santa Monica, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0715-9300","authenticated-orcid":false,"given":"Willi","family":"Menapace","sequence":"additional","affiliation":[{"name":"Snap Research, Santa Monica, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9252-1775","authenticated-orcid":false,"given":"Aliaksandr","family":"Siarohin","sequence":"additional","affiliation":[{"name":"Snap Research, Santa Monica, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3465-1592","authenticated-orcid":false,"given":"Sergey","family":"Tulyakov","sequence":"additional","affiliation":[{"name":"Snap Research, Santa Monica, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6777-7445","authenticated-orcid":false,"given":"Daniel","family":"Cohen-Or","sequence":"additional","affiliation":[{"name":"Snap Research, Palo Alto, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6028-6600","authenticated-orcid":false,"given":"Kfir","family":"Aberman","sequence":"additional","affiliation":[{"name":"Snap Research, Palo Alto, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,7,27]]},"reference":[{"key":"e_1_3_3_2_2_1","unstructured":"Michael\u00a0S Albergo and Eric Vanden-Eijnden. 2022. Building normalizing flows with stochastic interpolants. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2209.15571 (2022)."},{"key":"e_1_3_3_2_3_1","doi-asserted-by":"crossref","unstructured":"Omri Avrahami Kfir Aberman Ohad Fried Daniel Cohen-Or and Dani Lischinski. 2023. Break-A-Scene: Extracting Multiple Concepts from a Single Image. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2305.16311 (2023).","DOI":"10.1145\/3610548.3618154"},{"key":"e_1_3_3_2_4_1","unstructured":"Jianhong Bai Tianyu He Yuchi Wang Junliang Guo Haoji Hu Zuozhu Liu and Jiang Bian. 2024. UniEdit: A Unified Tuning-Free Framework for Video Motion and Appearance Editing. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2402.13185 (2024)."},{"key":"e_1_3_3_2_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/3680528.3687614"},{"key":"e_1_3_3_2_6_1","unstructured":"Xiuli Bi Jian Lu Bo Liu Xiaodong Cun Yong Zhang WeiSheng Li and Bin Xiao. 2024. CustomTTT: Motion and Appearance Customized Video Generation via Test-Time Training. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2412.15646 (2024)."},{"key":"e_1_3_3_2_7_1","unstructured":"Andreas Blattmann Tim Dockhorn Sumith Kulal Daniel Mendelevitch Maciej Kilian Dominik Lorenz Yam Levi Zion English Vikram Voleti Adam Letts et\u00a0al. 2023. Stable video diffusion: Scaling latent video diffusion models to large datasets. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2311.15127 (2023)."},{"key":"e_1_3_3_2_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.00727"},{"key":"e_1_3_3_2_9_1","unstructured":"Ting Chen and Lala Li. 2023. FIT: Far-reaching Interleaved Transformers. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2305.12689 (2023)."},{"key":"e_1_3_3_2_10_1","unstructured":"Tsai-Shien Chen Aliaksandr Siarohin Willi Menapace Yuwei Fang Ivan Skorokhodov Jun-Yan Zhu Kfir Aberman Ming-Hsuan Yang and Sergey Tulyakov. 2024. VideoAlchemy: Open-set Personalization in Video Generation. https:\/\/openreview.net\/forum?id=popKM1zAYa"},{"key":"e_1_3_3_2_11_1","unstructured":"Zuozhuo Dai Zhenghao Zhang Yao Yao Bingxue Qiu Siyu Zhu Long Qin and Weizhi Wang. 2023. AnimateAnything: Fine-Grained Open Domain Image Animation with Motion Guidance. arxiv:https:\/\/arXiv.org\/abs\/2311.12886\u00a0[cs.CV]"},{"key":"e_1_3_3_2_12_1","unstructured":"Google DeepMind. 2024. VEO2. https:\/\/deepmind.google\/technologies\/veo\/veo-2\/ (2024)."},{"key":"e_1_3_3_2_13_1","doi-asserted-by":"crossref","unstructured":"Jiankang Deng Jia Guo Jing Yang Niannan Xue Irene Kotsia and Stefanos Zafeiriou. 2022. ArcFace: Additive Angular Margin Loss for Deep Face Recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence 44 10 (oct 2022) 5962\u20135979. doi:10.1109\/tpami.2021.3087709","DOI":"10.1109\/TPAMI.2021.3087709"},{"key":"e_1_3_3_2_14_1","unstructured":"Rinon Gal Yuval Alaluf Yuval Atzmon Or Patashnik Amit\u00a0H Bermano Gal Chechik and Daniel Cohen-Or. 2022. An image is worth one word: Personalizing text-to-image generation using textual inversion. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2208.01618 (2022)."},{"key":"e_1_3_3_2_15_1","volume-title":"ICLR","author":"Gal Rinon","year":"2023","unstructured":"Rinon Gal, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit\u00a0H Bermano, Gal Chechik, and Daniel Cohen-Or. 2023. An image is worth one word: Personalizing text-to-image generation using textual inversion. In ICLR."},{"key":"e_1_3_3_2_16_1","unstructured":"Michal Geyer Omer Bar-Tal Shai Bagon and Tali Dekel. 2023. TokenFlow: Consistent Diffusion Features for Consistent Video Editing. arXiv preprint arxiv:https:\/\/arXiv.org\/abs\/2307.10373 (2023)."},{"key":"e_1_3_3_2_17_1","unstructured":"Litong Gong Yiran Zhu Weijie Li Xiaoyang Kang Biao Wang Tiezheng Ge and Bo Zheng. 2024. AtomoVideo: High Fidelity Image-to-Video Generation. arxiv:arXiv:2403.01800\u00a0[cs.CV]"},{"key":"e_1_3_3_2_18_1","unstructured":"Yuwei Guo Ceyuan Yang Anyi Rao Zhengyang Liang Yaohui Wang Yu Qiao Maneesh Agrawala Dahua Lin and Bo Dai. 2023. Animatediff: Animate your personalized text-to-image diffusion models without specific tuning. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2307.04725 (2023)."},{"key":"e_1_3_3_2_19_1","unstructured":"Yoav HaCohen Nisan Chiprut Benny Brazowski Daniel Shalem Dudu Moshe Eitan Richardson Eran Levin Guy Shiran Nir Zabari Ori Gordon Poriya Panet Sapir Weissbuch Victor Kulikov Yaki Bitterman Zeev Melumian and Ofir Bibi. 2024. LTX-Video: Realtime Video Latent Diffusion. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2501.00103 (2024)."},{"key":"e_1_3_3_2_20_1","unstructured":"Xuanhua He Quande Liu Shengju Qian Xin Wang Tao Hu Ke Cao Keyu Yan and Jie Zhang. 2024. ID-Animator: Zero-Shot Identity-Preserving Human Video Generation. arxiv:https:\/\/arXiv.org\/abs\/2404.15275\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2404.15275"},{"key":"e_1_3_3_2_21_1","unstructured":"Jonathan Ho William Chan Chitwan Saharia Jay Whang Ruiqi Gao Alexey Gritsenko Diederik\u00a0P Kingma Ben Poole Mohammad Norouzi David\u00a0J Fleet et\u00a0al. 2022. Imagen video: High definition video generation with diffusion models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2210.02303 (2022)."},{"key":"e_1_3_3_2_22_1","unstructured":"Wenyi Hong Ming Ding Wendi Zheng Xinghan Liu and Jie Tang. 2022. Cogvideo: Large-scale pretraining for text-to-video generation via transformers. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2205.15868 (2022)."},{"key":"e_1_3_3_2_23_1","unstructured":"Edward\u00a0J. Hu Yelong Shen Phillip Wallis Zeyuan Allen-Zhu Yuanzhi Li Shean Wang Lu Wang and Weizhu Chen. 2021. LoRA: Low-Rank Adaptation of Large Language Models. arxiv:https:\/\/arXiv.org\/abs\/2106.09685\u00a0[cs.CL] https:\/\/arxiv.org\/abs\/2106.09685"},{"key":"e_1_3_3_2_24_1","series-title":"(ICML\u201923)","volume-title":"Proceedings of the 40th International Conference on Machine Learning","author":"Jabri Allan","year":"2023","unstructured":"Allan Jabri, David\u00a0J. Fleet, and Ting Chen. 2023. Scalable adaptive computation for iterative generation. In Proceedings of the 40th International Conference on Machine Learning (Honolulu, Hawaii, USA) (ICML\u201923). JMLR.org, Article 594, 21\u00a0pages."},{"key":"e_1_3_3_2_25_1","unstructured":"Maxwell Jones Sheng-Yu Wang Nupur Kumari David Bau and Jun-Yan Zhu. 2024. Customizing Text-to-Image Models with a Single Image Pair. arxiv:https:\/\/arXiv.org\/abs\/2405.01536\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2405.01536"},{"key":"e_1_3_3_2_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.00622"},{"key":"e_1_3_3_2_27_1","unstructured":"Weijie Kong Qi Tian Zijian Zhang Rox Min Zuozhuo Dai Jin Zhou Jiangfeng Xiong Xin Li Bo Wu Jianwei Zhang et\u00a0al. 2024. HunyuanVideo: A Systematic Framework For Large Video Generative Models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2412.03603 (2024)."},{"key":"e_1_3_3_2_28_1","unstructured":"Feng Liang Bichen Wu Jialiang Wang Licheng Yu Kunpeng Li Yinan Zhao Ishan Misra Jia-Bin Huang Peizhao Zhang Peter Vajda et\u00a0al. 2023. FlowVid: Taming Imperfect Optical Flows for Consistent Video-to-Video Synthesis. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2312.17681 (2023)."},{"key":"e_1_3_3_2_29_1","unstructured":"Chang Liu Viraj Shah Aiyu Cui and Svetlana Lazebnik. 2024. UnZipLoRA: Separating Content and Style from a Single Image. arxiv:https:\/\/arXiv.org\/abs\/2412.04465\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2412.04465"},{"key":"e_1_3_3_2_30_1","unstructured":"Xingchao Liu Chengyue Gong and Qiang Liu. 2022. Flow straight and fast: Learning to generate and transfer data with rectified flow. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2209.03003 (2022)."},{"key":"e_1_3_3_2_31_1","unstructured":"Ze Ma Daquan Zhou Chun-Hsiao Yeh Xue-She Wang Xiuyu Li Huanrui Yang Zhen Dong Kurt Keutzer and Jiashi Feng. 2024. Magic-Me: Identity-Specific Video Customized Diffusion. arxiv:https:\/\/arXiv.org\/abs\/2402.09368\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2402.09368"},{"key":"e_1_3_3_2_32_1","first-page":"1634","volume-title":"Proceedings of the Asian Conference on Computer Vision","author":"Materzy\u0144ska Joanna","year":"2024","unstructured":"Joanna Materzy\u0144ska, Josef Sivic, Eli Shechtman, Antonio Torralba, Richard Zhang, and Bryan Russell. 2024. NewMove: Customizing text-to-video models with novel motions. In Proceedings of the Asian Conference on Computer Vision. 1634\u20131651."},{"key":"e_1_3_3_2_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.00672"},{"key":"e_1_3_3_2_34_1","unstructured":"Eyal Molad Eliahu Horwitz Dani Valevski Alex\u00a0Rav Acha Yossi Matias Yael Pritch Yaniv Leviathan and Yedid Hoshen. 2023. Dreamix: Video Diffusion Models are General Video Editors. arxiv:https:\/\/arXiv.org\/abs\/2302.01329\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2302.01329"},{"key":"e_1_3_3_2_35_1","unstructured":"OPENAI. 2024. SORA. https:\/\/openai.com\/sora\/ (2024)."},{"key":"e_1_3_3_2_36_1","doi-asserted-by":"crossref","unstructured":"William Peebles and Saining Xie. 2023. Scalable Diffusion Models with Transformers. IEEE 4172\u20134182.","DOI":"10.1109\/ICCV51070.2023.00387"},{"key":"e_1_3_3_2_37_1","unstructured":"Adam Polyak Amit Zohar Andrew Brown Andros Tjandra Animesh Sinha Ann Lee Apoorv Vyas Bowen Shi Chih-Yao Ma Ching-Yao Chuang et\u00a0al. 2024. Movie gen: A cast of media foundation models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2410.13720 (2024)."},{"key":"e_1_3_3_2_38_1","unstructured":"Chenyang Qi Xiaodong Cun Yong Zhang Chenyang Lei Xintao Wang Ying Shan and Qifeng Chen. 2023. FateZero: Fusing Attentions for Zero-shot Text-based Video Editing. arXiv:https:\/\/arXiv.org\/abs\/2303.09535 (2023)."},{"key":"e_1_3_3_2_39_1","unstructured":"Guocheng Qian Kuan-Chieh Wang Or Patashnik Negin Heravi Daniil Ostashev Sergey Tulyakov Daniel Cohen-Or and Kfir Aberman. 2024. Omni-ID: Holistic Identity Representation Designed for Generative Tasks. arxiv:https:\/\/arXiv.org\/abs\/2412.09694\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2412.09694"},{"key":"e_1_3_3_2_40_1","first-page":"8748","volume-title":"International conference on machine learning","author":"Radford Alec","year":"2021","unstructured":"Alec Radford, Jong\u00a0Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et\u00a0al. 2021. Learning transferable visual models from natural language supervision. In International conference on machine learning. PMLR, 8748\u20138763."},{"key":"e_1_3_3_2_41_1","unstructured":"Aditya Ramesh Prafulla Dhariwal Alex Nichol Casey Chu and Mark Chen. 2022. Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2204.06125 (2022)."},{"key":"e_1_3_3_2_42_1","unstructured":"Weiming Ren Harry Yang Ge Zhang Cong Wei Xinrun Du Stephen Huang and Wenhu Chen. 2024a. ConsistI2V: Enhancing Visual Consistency for Image-to-Video Generation. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2402.04324 (2024)."},{"key":"e_1_3_3_2_43_1","unstructured":"Yixuan Ren Yang Zhou Jimei Yang Jing Shi Difan Liu Feng Liu Mingi Kwon and Abhinav Shrivastava. 2024b. Customize-a-video: One-shot motion customization of text-to-video diffusion models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2402.14780 (2024)."},{"key":"e_1_3_3_2_44_1","doi-asserted-by":"crossref","unstructured":"Daniel Roich Ron Mokady Amit\u00a0H Bermano and Daniel Cohen-Or. 2022. Pivotal tuning for latent-based editing of real images. ACM Transactions on Graphics (TOG) 42 1 (2022) 1\u201313.","DOI":"10.1145\/3544777"},{"key":"e_1_3_3_2_45_1","doi-asserted-by":"crossref","unstructured":"Robin Rombach Andreas Blattmann Dominik Lorenz Patrick Esser and Bj\u00f6rn Ommer. 2022. High-resolution image synthesis with latent diffusion models. 10684\u201310695.","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"e_1_3_3_2_46_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"e_1_3_3_2_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.02155"},{"key":"e_1_3_3_2_48_1","unstructured":"Nataniel Ruiz Yuanzhen Li Varun Jampani Wei Wei Tingbo Hou Yael Pritch Neal Wadhwa Michael Rubinstein and Kfir Aberman. 2023b. Hyperdreambooth: Hypernetworks for fast personalization of text-to-image models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2307.06949 (2023)."},{"key":"e_1_3_3_2_49_1","unstructured":"Simo Ryu. 2023. DreamboothLoRA. https:\/\/github.com\/cloneofsimo\/lora"},{"key":"e_1_3_3_2_50_1","first-page":"36479","volume-title":"NIPS","author":"Saharia Chitwan","year":"2022","unstructured":"Chitwan Saharia, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily\u00a0L Denton, Kamyar Ghasemipour, Raphael Gontijo\u00a0Lopes, Burcu Karagol\u00a0Ayan, Tim Salimans, et\u00a0al. 2022. Photorealistic text-to-image diffusion models with deep language understanding. In NIPS. 36479\u201336494."},{"key":"e_1_3_3_2_51_1","unstructured":"Uriel Singer Adam Polyak Thomas Hayes Xi Yin Jie An Songyang Zhang Qiyuan Hu Harry Yang Oron Ashual Oran Gafni et\u00a0al. 2022. Make-a-video: Text-to-video generation without text-video data. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2209.14792 (2022)."},{"key":"e_1_3_3_2_52_1","unstructured":"Kuan-Chieh Wang Daniil Ostashev Yuwei Fang Sergey Tulyakov and Kfir Aberman. 2024. MoA: Mixture-of-Attention for Subject-Context Disentanglement in Personalized Image Generation. arxiv:https:\/\/arXiv.org\/abs\/2404.11565\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2404.11565"},{"key":"e_1_3_3_2_53_1","volume-title":"CVPR","author":"Wei Yujie","year":"2024","unstructured":"Yujie Wei, Shiwei Zhang, Zhiwu Qing, Hangjie Yuan, Zhiheng Liu, Yu Liu, Yingya Zhang, Jingren Zhou, and Hongming Shan. 2024. DreamVideo: Composing Your Dream Videos with Customized Subject and Motion. In CVPR."},{"key":"e_1_3_3_2_54_1","unstructured":"Jay\u00a0Zhangjie Wu Yixiao Ge Xintao Wang Weixian Lei Yuchao Gu Yufei Shi Wynne Hsu Ying Shan Xiaohu Qie and Mike\u00a0Zheng Shou. 2023. Tune-A-Video: One-Shot Tuning of Image Diffusion Models for Text-to-Video Generation. arxiv:https:\/\/arXiv.org\/abs\/2212.11565\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2212.11565"},{"key":"e_1_3_3_2_55_1","unstructured":"Zhuoyi Yang Jiayan Teng Wendi Zheng Ming Ding Shiyu Huang Jiazheng Xu Yuanming Yang Wenyi Hong Xiaohan Zhang Guanyu Feng et\u00a0al. 2024. CogVideoX: Text-to-Video Diffusion Models with An Expert Transformer. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2408.06072 (2024)."},{"key":"e_1_3_3_2_56_1","unstructured":"David\u00a0Junhao Zhang Dongxu Li Hung Le Mike\u00a0Zheng Shou Caiming Xiong and Doyen Sahoo. 2024. Moonshot: Towards Controllable Video Generation and Editing with Multimodal Conditions. arxiv:https:\/\/arXiv.org\/abs\/2401.01827\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2401.01827"},{"key":"e_1_3_3_2_57_1","unstructured":"Yuxin Zhang Fan Tang Nisha Huang Haibin Huang Chongyang Ma Weiming Dong and Changsheng Xu. 2023. MotionCrafter: One-Shot Motion Customization of Diffusion Models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2312.05288 (2023)."},{"key":"e_1_3_3_2_58_1","unstructured":"Rui Zhao Yuchao Gu Jay\u00a0Zhangjie Wu David\u00a0Junhao Zhang Jiawei Liu Weijia Wu Jussi Keppo and Mike\u00a0Zheng Shou. 2023. MotionDirector: Motion Customization of Text-to-Video Diffusion Models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2310.08465 (2023)."},{"key":"e_1_3_3_2_59_1","unstructured":"Yupeng Zhou Daquan Zhou Ming-Ming Cheng Jiashi Feng and Qibin Hou. 2024. StoryDiffusion: Consistent Self-Attention for Long-Range Image and Video Generation. arxiv:https:\/\/arXiv.org\/abs\/2405.01434\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2405.01434"}],"event":{"name":"SIGGRAPH Conference Papers '25: Special Interest Group on Computer Graphics and Interactive Techniques Conference Conference Papers","location":"Vancouver BC Canada","acronym":"SIGGRAPH Conference Papers '25","sponsor":["SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques"]},"container-title":["Proceedings of the Special Interest Group on Computer Graphics and Interactive Techniques Conference Conference Papers"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3721238.3730644","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,20]],"date-time":"2026-03-20T15:04:04Z","timestamp":1774019044000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3721238.3730644"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,27]]},"references-count":58,"alternative-id":["10.1145\/3721238.3730644","10.1145\/3721238"],"URL":"https:\/\/doi.org\/10.1145\/3721238.3730644","relation":{},"subject":[],"published":{"date-parts":[[2025,7,27]]},"assertion":[{"value":"2025-07-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}