{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,23]],"date-time":"2026-02-23T12:27:29Z","timestamp":1771849649071,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":76,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,12,10]],"date-time":"2023-12-10T00:00:00Z","timestamp":1702166400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,12,10]]},"DOI":"10.1145\/3610548.3618149","type":"proceedings-article","created":{"date-parts":[[2023,12,11]],"date-time":"2023-12-11T12:28:40Z","timestamp":1702297720000},"page":"1-12","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":13,"title":["iNVS: Repurposing Diffusion Inpainters for Novel View Synthesis"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0002-8347-4895","authenticated-orcid":false,"given":"Yash","family":"Kant","sequence":"first","affiliation":[{"name":"University of Toronto, Canada and Snap Inc., Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9252-1775","authenticated-orcid":false,"given":"Aliaksandr","family":"Siarohin","sequence":"additional","affiliation":[{"name":"Snap Inc., United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-4257-1636","authenticated-orcid":false,"given":"Michael","family":"Vasilkovsky","sequence":"additional","affiliation":[{"name":"Snap Inc., United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5149-5195","authenticated-orcid":false,"given":"Riza Alp","family":"Guler","sequence":"additional","affiliation":[{"name":"Snap Inc., United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3111-4134","authenticated-orcid":false,"given":"Jian","family":"Ren","sequence":"additional","affiliation":[{"name":"Snap Inc., United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3465-1592","authenticated-orcid":false,"given":"Sergey","family":"Tulyakov","sequence":"additional","affiliation":[{"name":"Snap Inc., United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6426-365X","authenticated-orcid":false,"given":"Igor","family":"Gilitschenski","sequence":"additional","affiliation":[{"name":"University of Toronto, Canada"}]}],"member":"320","published-online":{"date-parts":[[2023,12,11]]},"reference":[{"key":"e_1_3_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00580"},{"key":"e_1_3_2_2_2_1","volume-title":"Zoedepth: Zero-shot transfer by combining relative and metric depth. arXiv preprint arXiv:2302.12288","author":"Bhat Shariq\u00a0Farooq","year":"2023","unstructured":"Shariq\u00a0Farooq Bhat, Reiner Birkl, Diana Wofk, Peter Wonka, and Matthias M\u00fcller. 2023. Zoedepth: Zero-shot transfer by combining relative and metric depth. arXiv preprint arXiv:2302.12288 (2023)."},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00552"},{"key":"e_1_3_2_2_4_1","volume-title":"Persistent Nature: A Generative Model of Unbounded 3D Worlds. In CVPR.","author":"Chai Lucy","year":"2023","unstructured":"Lucy Chai, Richard Tucker, Zhengqi Li, Phillip Isola, and Noah Snavely. 2023. Persistent Nature: A Generative Model of Unbounded 3D Worlds. In CVPR."},{"key":"e_1_3_2_2_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00389"},{"key":"e_1_3_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01701"},{"key":"e_1_3_2_2_7_1","volume-title":"Nerv: Neural representations for videos. Advances in Neural Information Processing Systems","author":"Chen Hao","year":"2021","unstructured":"Hao Chen, Bo He, Hanyu Wang, Yixuan Ren, Ser\u00a0Nam Lim, and Abhinav Shrivastava. 2021. Nerv: Neural representations for videos. Advances in Neural Information Processing Systems (2021)."},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.02033"},{"key":"e_1_3_2_2_9_1","volume-title":"View Interpolation for Image Synthesis","author":"Chen Shenchang\u00a0Eric","unstructured":"Shenchang\u00a0Eric Chen and Lance Williams. 1993. View Interpolation for Image Synthesis. In Special Interest Group on Computer Graphics and Interactive Techniques."},{"key":"e_1_3_2_2_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00787"},{"key":"e_1_3_2_2_11_1","volume-title":"Modeling and Rendering Architecture from Photographs: A Hybrid Geometry- and Image-Based Approach","author":"Debevec E.","unstructured":"Paul\u00a0E. Debevec, Camillo\u00a0J. Taylor, and Jitendra Malik. 1996. Modeling and Rendering Architecture from Photographs: A Hybrid Geometry- and Image-Based Approach. In Special Interest Group on Computer Graphics and Interactive Techniques."},{"key":"e_1_3_2_2_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01263"},{"key":"e_1_3_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00431"},{"key":"e_1_3_2_2_14_1","volume-title":"Proceedings of the International Conference on Machine Learning.","author":"Denton Emily","year":"2018","unstructured":"Emily Denton and Rob Fergus. 2018. Stochastic video generation with a learned prior. In Proceedings of the International Conference on Machine Learning."},{"key":"e_1_3_2_2_15_1","volume-title":"Diffusion models beat gans on image synthesis. Advances in Neural Information Processing Systems","author":"Dhariwal Prafulla","year":"2021","unstructured":"Prafulla Dhariwal and Alexander Nichol. 2021. Diffusion models beat gans on image synthesis. Advances in Neural Information Processing Systems (2021)."},{"key":"e_1_3_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA46639.2022.9811809"},{"key":"e_1_3_2_2_17_1","volume-title":"Unsupervised learning for physical interaction through video prediction. Advances in Neural Information Processing Systems","author":"Finn Chelsea","year":"2016","unstructured":"Chelsea Finn, Ian Goodfellow, and Sergey Levine. 2016. Unsupervised learning for physical interaction through video prediction. Advances in Neural Information Processing Systems (2016)."},{"key":"e_1_3_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00566"},{"key":"e_1_3_2_2_19_1","volume-title":"Proceedings of the International Conference on Machine Learning.","author":"Gu Jiatao","year":"2023","unstructured":"Jiatao Gu, Alex Trevithick, Kai-En Lin, Joshua\u00a0M Susskind, Christian Theobalt, Lingjie Liu, and Ravi Ramamoorthi. 2023. Nerfdiff: Single-image view synthesis with nerf-guided distillation from 3d-aware diffusion. In Proceedings of the International Conference on Machine Learning."},{"key":"e_1_3_2_2_20_1","volume-title":"Learning to decompose and disentangle representations for video prediction. Advances in Neural Information Processing Systems","author":"Hsieh Jun-Ting","year":"2018","unstructured":"Jun-Ting Hsieh, Bingbin Liu, De-An Huang, Li\u00a0F Fei-Fei, and Juan\u00a0Carlos Niebles. 2018. Learning to decompose and disentangle representations for video prediction. Advances in Neural Information Processing Systems (2018)."},{"key":"e_1_3_2_2_21_1","volume-title":"Edmond Shu-Lim Ho, and Adrian Munteanu","author":"Hu Pengpeng","year":"2021","unstructured":"Pengpeng Hu, Edmond Shu-Lim Ho, and Adrian Munteanu. 2021a. 3DBodyNet: fast reconstruction of 3D animatable human body shape from a single commodity depth camera. IEEE Transactions on Multimedia (2021)."},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01230"},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2018.2845697"},{"key":"e_1_3_2_2_24_1","volume-title":"Shap-e: Generating conditional 3d implicit functions. arXiv preprint arXiv:2305.02463","author":"Jun Heewoo","year":"2023","unstructured":"Heewoo Jun and Alex Nichol. 2023. Shap-e: Generating conditional 3d implicit functions. arXiv preprint arXiv:2305.02463 (2023)."},{"key":"e_1_3_2_2_25_1","volume-title":"3D face reconstruction from a single image using a single reference face shape","author":"Kemelmacher-Shlizerman Ira","year":"2010","unstructured":"Ira Kemelmacher-Shlizerman and Ronen Basri. 2010. 3D face reconstruction from a single image using a single reference face shape. IEEE Transactions on Pattern Analysis and Machine Intelligence (2010)."},{"key":"e_1_3_2_2_26_1","volume-title":"Pathdreamer: A world model for indoor navigation. ICCV.","author":"Koh Jing\u00a0Yu","year":"2021","unstructured":"Jing\u00a0Yu Koh, Honglak Lee, Yinfei Yang, Jason Baldridge, and Peter Anderson. 2021. Pathdreamer: A world model for indoor navigation. ICCV."},{"key":"e_1_3_2_2_27_1","volume-title":"NeROIC: Neural Rendering of Objects from Online Image Collections","author":"Kuang Zhengfei","unstructured":"Zhengfei Kuang, Kyle Olszewski, Menglei Chai, Zeng Huang, Panos Achlioptas, and Sergey Tulyakov. 2022. NeROIC: Neural Rendering of Objects from Online Image Collections. In Special Interest Group on Computer Graphics and Interactive Techniques."},{"key":"e_1_3_2_2_28_1","unstructured":"Lambda Labs. 2023. Stable Diffusion Image Variations. https:\/\/huggingface.co\/spaces\/lambdalabs\/stable-diffusion-image-variations. Accessed on 2023-05-22."},{"key":"e_1_3_2_2_29_1","volume-title":"Proceedings of the International Conference on Learning Representations.","author":"Lee Wonkwang","year":"2021","unstructured":"Wonkwang Lee, Whie Jung, Han Zhang, Ting Chen, Jing\u00a0Yu Koh, Thomas Huang, Hyungsuk Yoon, Honglak Lee, and Seunghoon Hong. 2021. Revisiting hierarchical approach for persistent long-term video prediction. In Proceedings of the International Conference on Learning Representations."},{"key":"e_1_3_2_2_30_1","volume-title":"Generative Image Dynamics. arXiv preprint arXiv:2309.07906","author":"Li Zhengqi","year":"2023","unstructured":"Zhengqi Li, Richard Tucker, Noah Snavely, and Aleksander Holynski. 2023. Generative Image Dynamics. arXiv preprint arXiv:2309.07906 (2023)."},{"key":"e_1_3_2_2_31_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19769-7_30"},{"key":"e_1_3_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19769-7_30"},{"key":"e_1_3_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00837"},{"key":"e_1_3_2_2_34_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00037"},{"key":"e_1_3_2_2_35_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01419"},{"key":"e_1_3_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00853"},{"key":"e_1_3_2_2_37_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01117"},{"key":"e_1_3_2_2_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00365"},{"key":"e_1_3_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00713"},{"key":"e_1_3_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58452-8_24"},{"key":"e_1_3_2_2_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01633"},{"key":"e_1_3_2_2_42_1","volume-title":"Point-e: A system for generating 3d point clouds from complex prompts. arXiv preprint arXiv:2212.08751","author":"Nichol Alex","year":"2022","unstructured":"Alex Nichol, Heewoo Jun, Prafulla Dhariwal, Pamela Mishkin, and Mark Chen. 2022. Point-e: A system for generating 3d point clouds from complex prompts. arXiv preprint arXiv:2212.08751 (2022)."},{"key":"e_1_3_2_2_43_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00548"},{"key":"e_1_3_2_2_44_1","volume-title":"Proceedings of the International Conference on Learning Representations.","author":"Poole Ben","year":"2023","unstructured":"Ben Poole, Ajay Jain, Jonathan\u00a0T. Barron, and Ben Mildenhall. 2023. DreamFusion: Text-to-3D using 2D Diffusion. In Proceedings of the International Conference on Learning Representations."},{"key":"e_1_3_2_2_45_1","volume-title":"Proceedings of the International Conference on Machine Learning.","author":"Radford Alec","year":"2021","unstructured":"Alec Radford, Jong\u00a0Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, 2021. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning."},{"key":"e_1_3_2_2_46_1","volume-title":"DreamBooth3D: Subject-Driven Text-to-3D Generation. arXiv preprint arXiv:2303.13508","author":"Raj Amit","year":"2023","unstructured":"Amit Raj, Srinivas Kaza, Ben Poole, Michael Niemeyer, Nataniel Ruiz, Ben Mildenhall, Shiran Zada, Kfir Aberman, Michael Rubinstein, Jonathan Barron, 2023. DreamBooth3D: Subject-Driven Text-to-3D Generation. arXiv preprint arXiv:2303.13508 (2023)."},{"key":"e_1_3_2_2_47_1","volume-title":"Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125","author":"Ramesh Aditya","year":"2022","unstructured":"Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. 2022. Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125 (2022)."},{"key":"e_1_3_2_2_48_1","volume-title":"Proceedings of the International Conference on Machine Learning.","author":"Ramesh Aditya","year":"2021","unstructured":"Aditya Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Radford, Mark Chen, and Ilya Sutskever. 2021. Zero-shot text-to-image generation. In Proceedings of the International Conference on Machine Learning."},{"key":"e_1_3_2_2_49_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01072"},{"key":"e_1_3_2_2_50_1","volume-title":"TEXTure: Text-Guided Texturing of 3D Shapes","author":"Richardson Elad","unstructured":"Elad Richardson, Gal Metzer, Yuval Alaluf, Raja Giryes, and Daniel Cohen-Or. 2023. TEXTure: Text-Guided Texturing of 3D Shapes. In Special Interest Group on Computer Graphics and Interactive Techniques."},{"key":"e_1_3_2_2_51_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"e_1_3_2_2_52_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"e_1_3_2_2_53_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.02155"},{"key":"e_1_3_2_2_54_1","volume-title":"Photorealistic text-to-image diffusion models with deep language understanding. Advances in Neural Information Processing Systems","author":"Saharia Chitwan","year":"2022","unstructured":"Chitwan Saharia, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily\u00a0L Denton, Kamyar Ghasemipour, Raphael Gontijo\u00a0Lopes, Burcu Karagol\u00a0Ayan, Tim Salimans, 2022. Photorealistic text-to-image diffusion models with deep language understanding. Advances in Neural Information Processing Systems (2022)."},{"key":"e_1_3_2_2_55_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00613"},{"key":"e_1_3_2_2_56_1","volume-title":"Laion-5b: An open large-scale dataset for training next generation image-text models. Advances in Neural Information Processing Systems","author":"Schuhmann Christoph","year":"2022","unstructured":"Christoph Schuhmann, Romain Beaumont, Richard Vencu, Cade Gordon, Ross Wightman, Mehdi Cherti, Theo Coombes, Aarush Katta, Clayton Mullis, Mitchell Wortsman, 2022. Laion-5b: An open large-scale dataset for training next generation image-text models. Advances in Neural Information Processing Systems (2022)."},{"key":"e_1_3_2_2_57_1","volume-title":"Anything-3d: Towards single-view anything reconstruction in the wild. arXiv preprint arXiv:2304.10261","author":"Shen Qiuhong","year":"2023","unstructured":"Qiuhong Shen, Xingyi Yang, and Xinchao Wang. 2023. Anything-3d: Towards single-view anything reconstruction in the wild. arXiv preprint arXiv:2304.10261 (2023)."},{"key":"e_1_3_2_2_58_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00805"},{"key":"e_1_3_2_2_59_1","volume-title":"First order motion model for image animation. Advances in Neural Information Processing Systems","author":"Siarohin Aliaksandr","year":"2019","unstructured":"Aliaksandr Siarohin, St\u00e9phane Lathuili\u00e8re, Sergey Tulyakov, Elisa Ricci, and Nicu Sebe. 2019a. First order motion model for image animation. Advances in Neural Information Processing Systems (2019)."},{"key":"e_1_3_2_2_60_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00248"},{"key":"e_1_3_2_2_61_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.02086"},{"key":"e_1_3_2_2_62_1","volume-title":"Proceedings of the European Conference on Computer Vision Workshop","author":"Tremblay Jonathan","year":"2022","unstructured":"Jonathan Tremblay, Moustafa Meshry, Alex Evans, Jan Kautz, Alexander Keller, Sameh Khamis, Charles Loop, Nathan Morrical, Koki Nagano, Towaki Takikawa, and Stan Birchfield. 2022. RTMV: A Ray-Traced Multi-View Synthetic Dataset for Novel View Synthesis. Proceedings of the European Conference on Computer Vision Workshop (2022)."},{"key":"e_1_3_2_2_63_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00063"},{"key":"e_1_3_2_2_64_1","volume-title":"Proceedings of the International Conference on Learning Representations.","author":"Villegas Ruben","year":"2017","unstructured":"Ruben Villegas, Jimei Yang, Seunghoon Hong, Xunyu Lin, and Honglak Lee. 2017. Decomposing motion and content for natural video sequence prediction. In Proceedings of the International Conference on Learning Representations."},{"key":"e_1_3_2_2_65_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.319"},{"key":"e_1_3_2_2_66_1","volume-title":"PredRNN: Recurrent Neural Networks for Predictive Learning Using Spatiotemporal LSTMs. Advances in Neural Information Processing Systems","author":"Wang Yunbo","year":"2017","unstructured":"Yunbo Wang, Mingsheng Long, Jianmin Wang, Zhifeng Gao, and Philip\u00a0S Yu. 2017. PredRNN: Recurrent Neural Networks for Predictive Learning Using Spatiotemporal LSTMs. Advances in Neural Information Processing Systems (2017)."},{"key":"e_1_3_2_2_67_1","doi-asserted-by":"crossref","unstructured":"Zhou Wang A.C. Bovik H.R. Sheikh and E.P. Simoncelli. 2004. Image quality assessment: from error visibility to structural similarity. IEEE Transactions on Image Processing (2004).","DOI":"10.1109\/TIP.2003.819861"},{"key":"e_1_3_2_2_68_1","volume-title":"NeRF \u2212 \u2212: Neural Radiance Fields Without Known Camera Parameters. arXiv preprint arXiv:2102.07064","author":"Wang Zirui","year":"2021","unstructured":"Zirui Wang, Shangzhe Wu, Weidi Xie, Min Chen, and Victor\u00a0Adrian Prisacariu. 2021. NeRF \u2212 \u2212: Neural Radiance Fields Without Known Camera Parameters. arXiv preprint arXiv:2102.07064 (2021)."},{"key":"e_1_3_2_2_69_1","volume-title":"Proceedings of the International Conference on Learning Representations.","author":"Watson Daniel","year":"2022","unstructured":"Daniel Watson, William Chan, Ricardo\u00a0Martin Brualla, Jonathan Ho, Andrea Tagliasacchi, and Mohammad Norouzi. 2022. Novel View Synthesis with Diffusion Models. In Proceedings of the International Conference on Learning Representations."},{"key":"e_1_3_2_2_70_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00749"},{"key":"e_1_3_2_2_71_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00057"},{"key":"e_1_3_2_2_72_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.01045"},{"key":"e_1_3_2_2_73_1","volume-title":"Proceedings of the International Conference on Learning Representations.","author":"Yu Sihyun","year":"2022","unstructured":"Sihyun Yu, Jihoon Tack, Sangwoo Mo, Hyunsu Kim, Junho Kim, Jung-Woo Ha, and Jinwoo Shin. 2022. Generating videos with dynamics-aware implicit generative adversarial networks. In Proceedings of the International Conference on Learning Representations."},{"key":"e_1_3_2_2_74_1","volume-title":"Analyzing and improving neural radiance fields. arXiv preprint arXiv:2010.07492","author":"Zhang Kai","year":"2020","unstructured":"Kai Zhang, Gernot Riegler, Noah Snavely, and Vladlen Koltun. 2020. Nerf++: Analyzing and improving neural radiance fields. arXiv preprint arXiv:2010.07492 (2020)."},{"key":"e_1_3_2_2_75_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00068"},{"key":"e_1_3_2_2_76_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01211"}],"event":{"name":"SA '23: SIGGRAPH Asia 2023","location":"Sydney NSW Australia","acronym":"SA '23","sponsor":["SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques"]},"container-title":["SIGGRAPH Asia 2023 Conference Papers"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3610548.3618149","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3610548.3618149","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T09:32:25Z","timestamp":1755768745000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3610548.3618149"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,10]]},"references-count":76,"alternative-id":["10.1145\/3610548.3618149","10.1145\/3610548"],"URL":"https:\/\/doi.org\/10.1145\/3610548.3618149","relation":{},"subject":[],"published":{"date-parts":[[2023,12,10]]},"assertion":[{"value":"2023-12-11","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}