{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T19:45:44Z","timestamp":1776887144081,"version":"3.51.2"},"publisher-location":"New York, NY, USA","reference-count":57,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,12,10]],"date-time":"2023-12-10T00:00:00Z","timestamp":1702166400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61976250"],"award-info":[{"award-number":["61976250"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Guangdong Basic and Applied Basic Research Foundation","award":["2020B1515020048"],"award-info":[{"award-number":["2020B1515020048"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,12,10]]},"DOI":"10.1145\/3610548.3618190","type":"proceedings-article","created":{"date-parts":[[2023,12,11]],"date-time":"2023-12-11T12:28:40Z","timestamp":1702297720000},"page":"1-10","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":98,"title":["DreamEditor: Text-Driven 3D Scene Editing with Neural Fields"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4378-0890","authenticated-orcid":false,"given":"Jingyu","family":"Zhuang","sequence":"first","affiliation":[{"name":"Sun Yat-sen University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9315-3780","authenticated-orcid":false,"given":"Chen","family":"Wang","sequence":"additional","affiliation":[{"name":"University of Pennsylvania, United States of America and Tsinghua University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2248-3755","authenticated-orcid":false,"given":"Liang","family":"Lin","sequence":"additional","affiliation":[{"name":"Sun Yat-sen University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4301-1474","authenticated-orcid":false,"given":"Lingjie","family":"Liu","sequence":"additional","affiliation":[{"name":"University of Pennsylvania, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4805-0926","authenticated-orcid":false,"given":"Guanbin","family":"Li","sequence":"additional","affiliation":[{"name":"Sun Yat-sen University, China"}]}],"member":"320","published-online":{"date-parts":[[2023,12,11]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01767"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.02004"},{"key":"e_1_3_2_1_3_1","volume-title":"Instructpix2pix: Learning to follow image editing instructions. arXiv preprint arXiv:2211.09800","author":"Brooks Tim","year":"2022","unstructured":"Tim Brooks, Aleksander Holynski, and Alexei\u00a0A Efros. 2022. Instructpix2pix: Learning to follow image editing instructions. arXiv preprint arXiv:2211.09800 (2022)."},{"key":"e_1_3_2_1_4_1","volume-title":"Animatable neural radiance fields from monocular rgb videos. arXiv preprint arXiv:2106.13629","author":"Chen Jianchuan","year":"2021","unstructured":"Jianchuan Chen, Ying Zhang, Di Kang, Xuefei Zhe, Linchao Bao, Xu Jia, and Huchuan Lu. 2021. Animatable neural radiance fields from monocular rgb videos. arXiv preprint arXiv:2106.13629 (2021)."},{"key":"e_1_3_2_1_5_1","volume-title":"Fantasia3D: Disentangling Geometry and Appearance for High-quality Text-to-3D Content Creation. arXiv preprint arXiv:2303.13873","author":"Chen Rui","year":"2023","unstructured":"Rui Chen, Yongwei Chen, Ningxin Jiao, and Kui Jia. 2023. Fantasia3D: Disentangling Geometry and Appearance for High-quality Text-to-3D Content Creation. arXiv preprint arXiv:2303.13873 (2023)."},{"key":"e_1_3_2_1_6_1","volume-title":"Tango: Text-driven photorealistic and robust 3d stylization via lighting decomposition. arXiv preprint arXiv:2210.11277","author":"Chen Yongwei","year":"2022","unstructured":"Yongwei Chen, Rui Chen, Jiabao Lei, Yabin Zhang, and Kui Jia. 2022. Tango: Text-driven photorealistic and robust 3d stylization via lighting decomposition. arXiv preprint arXiv:2210.11277 (2022)."},{"key":"e_1_3_2_1_7_1","volume-title":"Diffedit: Diffusion-based semantic image editing with mask guidance. arXiv preprint arXiv:2210.11427","author":"Couairon Guillaume","year":"2022","unstructured":"Guillaume Couairon, Jakob Verbeek, Holger Schwenk, and Matthieu Cord. 2022. Diffedit: Diffusion-based semantic image editing with mask guidance. arXiv preprint arXiv:2210.11427 (2022)."},{"key":"e_1_3_2_1_8_1","volume-title":"An image is worth one word: Personalizing text-to-image generation using textual inversion. arXiv preprint arXiv:2208.01618","author":"Gal Rinon","year":"2022","unstructured":"Rinon Gal, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit\u00a0H Bermano, Gal Chechik, and Daniel Cohen-Or. 2022a. An image is worth one word: Personalizing text-to-image generation using textual inversion. arXiv preprint arXiv:2208.01618 (2022)."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3528223.3530164"},{"key":"e_1_3_2_1_10_1","volume-title":"TextDeformer: Geometry Manipulation using Text Guidance. arXiv preprint arXiv:2304.13348","author":"Gao William","year":"2023","unstructured":"William Gao, Noam Aigerman, Thibault Groueix, Vladimir\u00a0G Kim, and Rana Hanocka. 2023. TextDeformer: Geometry Manipulation using Text Guidance. arXiv preprint arXiv:2304.13348 (2023)."},{"key":"e_1_3_2_1_11_1","volume-title":"Implicit geometric regularization for learning shapes. arXiv preprint arXiv:2002.10099","author":"Gropp Amos","year":"2020","unstructured":"Amos Gropp, Lior Yariv, Niv Haim, Matan Atzmon, and Yaron Lipman. 2020. Implicit geometric regularization for learning shapes. arXiv preprint arXiv:2002.10099 (2020)."},{"key":"e_1_3_2_1_12_1","volume-title":"Instruct-NeRF2NeRF: Editing 3D Scenes with Instructions. arXiv preprint arXiv:2303.12789","author":"Haque Ayaan","year":"2023","unstructured":"Ayaan Haque, Matthew Tancik, Alexei\u00a0A Efros, Aleksander Holynski, and Angjoo Kanazawa. 2023. Instruct-NeRF2NeRF: Editing 3D Scenes with Instructions. arXiv preprint arXiv:2303.12789 (2023)."},{"key":"e_1_3_2_1_13_1","volume-title":"Prompt-to-prompt image editing with cross attention control. arXiv preprint arXiv:2208.01626","author":"Hertz Amir","year":"2022","unstructured":"Amir Hertz, Ron Mokady, Jay Tenenbaum, Kfir Aberman, Yael Pritch, and Daniel Cohen-Or. 2022. Prompt-to-prompt image editing with cross attention control. arXiv preprint arXiv:2208.01626 (2022)."},{"key":"e_1_3_2_1_14_1","volume-title":"Denoising diffusion probabilistic models. NeurIPS 2020 33","author":"Ho Jonathan","year":"2020","unstructured":"Jonathan Ho, Ajay Jain, and Pieter Abbeel. 2020. Denoising diffusion probabilistic models. NeurIPS 2020 33 (2020), 6840\u20136851."},{"key":"e_1_3_2_1_15_1","volume-title":"CVPR","author":"Jain Ajay","year":"2022","unstructured":"Ajay Jain, Ben Mildenhall, Jonathan\u00a0T Barron, and et al.2022. Zero-shot text-guided object generation with dream fields. In CVPR 2022. 867\u2013876."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.59"},{"key":"e_1_3_2_1_17_1","volume-title":"Imagic: Text-based real image editing with diffusion models. arXiv preprint arXiv:2210.09276","author":"Kawar Bahjat","year":"2022","unstructured":"Bahjat Kawar, Shiran Zada, Oran Lang, Omer Tov, Huiwen Chang, Tali Dekel, Inbar Mosseri, and Michal Irani. 2022. Imagic: Text-based real image editing with diffusion models. arXiv preprint arXiv:2210.09276 (2022)."},{"key":"e_1_3_2_1_18_1","volume-title":"Decomposing nerf for editing via feature field distillation. arXiv preprint arXiv:2205.15585","author":"Kobayashi Sosuke","year":"2022","unstructured":"Sosuke Kobayashi, Eiichi Matsumoto, and Vincent Sitzmann. 2022. Decomposing nerf for editing via feature field distillation. arXiv preprint arXiv:2205.15585 (2022)."},{"key":"e_1_3_2_1_19_1","volume-title":"ClimateNeRF: Physically-based Neural Rendering for Extreme Climate Synthesis. arXiv e-prints","author":"Li Yuan","year":"2022","unstructured":"Yuan Li, Zhi-Hao Lin, David Forsyth, Jia-Bin Huang, and Shenlong Wang. 2022. ClimateNeRF: Physically-based Neural Rendering for Extreme Climate Synthesis. arXiv e-prints (2022), arXiv\u20132211."},{"key":"e_1_3_2_1_20_1","volume-title":"Magic3D: High-Resolution Text-to-3D Content Creation. arXiv preprint arXiv:2211.10440","author":"Lin Chen-Hsuan","year":"2022","unstructured":"Chen-Hsuan Lin, Jun Gao, Luming Tang, Towaki Takikawa, Xiaohui Zeng, Xun Huang, Karsten Kreis, Sanja Fidler, Ming-Yu Liu, and Tsung-Yi Lin. 2022. Magic3D: High-Resolution Text-to-3D Content Creation. arXiv preprint arXiv:2211.10440 (2022)."},{"key":"e_1_3_2_1_21_1","volume-title":"NeRF-In: Free-form NeRF inpainting with RGB-D priors. arXiv preprint arXiv:2206.04901","author":"Liu Hao-Kang","year":"2022","unstructured":"Hao-Kang Liu, I Shen, Bing-Yu Chen, 2022. NeRF-In: Free-form NeRF inpainting with RGB-D priors. arXiv preprint arXiv:2206.04901 (2022)."},{"key":"e_1_3_2_1_22_1","volume-title":"2020. Neural sparse voxel fields. NeurIPS 2020 33","author":"Liu Lingjie","year":"2020","unstructured":"Lingjie Liu, Jiatao Gu, Kyaw Zaw\u00a0Lin, and et al.2020. Neural sparse voxel fields. NeurIPS 2020 33 (2020), 15651\u201315663."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00572"},{"key":"e_1_3_2_1_24_1","volume-title":"Marching cubes: A high resolution 3D surface construction algorithm. ACM siggraph computer graphics 21, 4","author":"Lorensen E","year":"1987","unstructured":"William\u00a0E Lorensen and Harvey\u00a0E Cline. 1987. Marching cubes: A high resolution 3D surface construction algorithm. ACM siggraph computer graphics 21, 4 (1987), 163\u2013169."},{"key":"e_1_3_2_1_25_1","volume-title":"Latent-NeRF for Shape-Guided Generation of 3D Shapes and Textures. arXiv preprint arXiv:2211.07600","author":"Metzer Gal","year":"2022","unstructured":"Gal Metzer, Elad Richardson, Or Patashnik, Raja Giryes, and Daniel Cohen-Or. 2022. Latent-NeRF for Shape-Guided Generation of 3D Shapes and Textures. arXiv preprint arXiv:2211.07600 (2022)."},{"key":"e_1_3_2_1_26_1","volume-title":"CVPR","author":"Michel Oscar","year":"2022","unstructured":"Oscar Michel, Roi Bar-On, Richard Liu, and et al.2022. Text2mesh: Text-driven neural stylization for meshes. In CVPR 2022. 13492\u201313502."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3503250"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3550469.3555392"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3528223.3530127"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00571"},{"key":"e_1_3_2_1_31_1","volume-title":"CVPR 2022. 1841","author":"Ost Julian","year":"1842","unstructured":"Julian Ost, Issam Laradji, Alejandro Newell, and et al.2022. Neural point light fields. In CVPR 2022. 18419\u201318429."},{"key":"e_1_3_2_1_32_1","volume-title":"CVPR","author":"Peng Sida","year":"2021","unstructured":"Sida Peng, Yuanqing Zhang, Yinghao Xu, and et al.2021. Neural body: Implicit neural representations with structured latent codes for novel view synthesis of dynamic humans. In CVPR 2021. 9054\u20139063."},{"key":"e_1_3_2_1_33_1","volume-title":"Dreamfusion: Text-to-3d using 2d diffusion. arXiv preprint arXiv:2209.14988","author":"Poole Ben","year":"2022","unstructured":"Ben Poole, Ajay Jain, Jonathan\u00a0T Barron, and Ben Mildenhall. 2022. Dreamfusion: Text-to-3d using 2d diffusion. arXiv preprint arXiv:2209.14988 (2022)."},{"key":"e_1_3_2_1_34_1","volume-title":"CVPR","author":"Qi R","year":"2017","unstructured":"Charles\u00a0R Qi, Hao Su, Kaichun Mo, and Leonidas\u00a0J Guibas. 2017. Pointnet: Deep learning on point sets for 3d classification and segmentation. In CVPR 2017. 652\u2013660."},{"key":"e_1_3_2_1_35_1","volume-title":"DreamBooth3D: Subject-Driven Text-to-3D Generation. arXiv preprint arXiv:2303.13508","author":"Raj Amit","year":"2023","unstructured":"Amit Raj, Srinivas Kaza, Ben Poole, Michael Niemeyer, Nataniel Ruiz, Ben Mildenhall, Shiran Zada, Kfir Aberman, Michael Rubinstein, Jonathan Barron, 2023. DreamBooth3D: Subject-Driven Text-to-3D Generation. arXiv preprint arXiv:2303.13508 (2023)."},{"key":"e_1_3_2_1_36_1","volume-title":"Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125","author":"Ramesh Aditya","year":"2022","unstructured":"Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. 2022. Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125 (2022)."},{"key":"e_1_3_2_1_37_1","volume-title":"ICCV","author":"Reizenstein Jeremy","year":"2021","unstructured":"Jeremy Reizenstein, Roman Shapovalov, Philipp Henzler, and et al.2021. Common objects in 3d: Large-scale learning and evaluation of real-life 3d category reconstruction. In ICCV 2021. 10901\u201310911."},{"key":"e_1_3_2_1_38_1","volume-title":"Texture: Text-guided texturing of 3d shapes. arXiv preprint arXiv:2302.01721","author":"Richardson Elad","year":"2023","unstructured":"Elad Richardson, Gal Metzer, Yuval Alaluf, Raja Giryes, and Daniel Cohen-Or. 2023. Texture: Text-guided texturing of 3d shapes. arXiv preprint arXiv:2302.01721 (2023)."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"e_1_3_2_1_40_1","volume-title":"Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation. arXiv preprint arXiv:2208.12242","author":"Ruiz Nataniel","year":"2022","unstructured":"Nataniel Ruiz, Yuanzhen Li, Varun Jampani, Yael Pritch, Michael Rubinstein, and Kfir Aberman. 2022. Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation. arXiv preprint arXiv:2208.12242 (2022)."},{"key":"e_1_3_2_1_41_1","volume-title":"Photorealistic text-to-image diffusion models with deep language understanding. NeurIPS 2022 35","author":"Saharia Chitwan","year":"2022","unstructured":"Chitwan Saharia, William Chan, and Saurabh et\u00a0al. Saxena. 2022. Photorealistic text-to-image diffusion models with deep language understanding. NeurIPS 2022 35 (2022), 36479\u201336494."},{"key":"e_1_3_2_1_42_1","volume-title":"Vox-E: Text-guided Voxel Editing of 3D Objects. arXiv preprint arXiv:2303.12048","author":"Sella Etai","year":"2023","unstructured":"Etai Sella, Gal Fiebelman, Peter Hedman, and Hadar Averbuch-Elor. 2023. Vox-E: Text-guided Voxel Editing of 3D Objects. arXiv preprint arXiv:2303.12048 (2023)."},{"key":"e_1_3_2_1_43_1","volume-title":"ACCV","author":"Shen Tianwei","year":"2018","unstructured":"Tianwei Shen, Zixin Luo, Lei Zhou, and et al.2018. Matchable image retrieval by learning from surface reconstruction. In ACCV 2018. Springer, 415\u2013431."},{"key":"e_1_3_2_1_44_1","volume-title":"Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502","author":"Song Jiaming","year":"2020","unstructured":"Jiaming Song, Chenlin Meng, and Stefano Ermon. 2020. Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502 (2020)."},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"crossref","unstructured":"Robert\u00a0W Sumner Johannes Schmid and Mark Pauly. 2007. Embedded deformation for shape manipulation. In ACM siggraph 2007 papers. 80\u2013es.","DOI":"10.1145\/1275808.1276478"},{"key":"e_1_3_2_1_46_1","volume-title":"2017. Attention is all you need. NeurIPS 2017 30","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, and et al.2017. Attention is all you need. NeurIPS 2017 30 (2017)."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00381"},{"key":"e_1_3_2_1_48_1","volume-title":"Nerf-art: Text-driven neural radiance fields stylization","author":"Wang Can","year":"2023","unstructured":"Can Wang, Ruixiang Jiang, Menglei Chai, Mingming He, Dongdong Chen, and Jing Liao. 2023. Nerf-art: Text-driven neural radiance fields stylization. IEEE Transactions on Visualization and Computer Graphics (2023)."},{"key":"e_1_3_2_1_49_1","volume-title":"ACM MM","author":"Wang Chen","year":"2022","unstructured":"Chen Wang, Xian Wu, Yuan-Chen Guo, and et al.2022c. NeRF-SR: High Quality Neural Radiance Fields using Supersampling. In ACM MM 2022. 6445\u20136454."},{"key":"e_1_3_2_1_50_1","volume-title":"Score Jacobian Chaining: Lifting Pretrained 2D Diffusion Models for 3D Generation. arXiv preprint arXiv:2212.00774","author":"Wang Haochen","year":"2022","unstructured":"Haochen Wang, Xiaodan Du, Jiahao Li, Raymond\u00a0A Yeh, and Greg Shakhnarovich. 2022b. Score Jacobian Chaining: Lifting Pretrained 2D Diffusion Models for 3D Generation. arXiv preprint arXiv:2212.00774 (2022)."},{"key":"e_1_3_2_1_51_1","volume-title":"Neus: Learning neural implicit surfaces by","author":"Wang Peng","year":"2021","unstructured":"Peng Wang, Lingjie Liu, Yuan Liu, Christian Theobalt, Taku Komura, and Wenping Wang. 2021. Neus: Learning neural implicit surfaces by volume rendering for multi-view reconstruction. arXiv preprint arXiv:2106.10689 (2021)."},{"key":"e_1_3_2_1_52_1","volume-title":"CVPR","author":"Xiang Fanbo","year":"2021","unstructured":"Fanbo Xiang, Zexiang Xu, Milos Hasan, and et al.2021. Neutex: Neural texture mapping for volumetric neural rendering. In CVPR 2021. 7119\u20137128."},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19827-4_10"},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19787-1_34"},{"key":"e_1_3_2_1_55_1","volume-title":"BlendedMVS: A Large-scale Dataset for Generalized Multi-view Stereo Networks. In CVPR","author":"Yao Yao","year":"2020","unstructured":"Yao Yao, Zixin Luo, Shiwei Li, and et al.2020. BlendedMVS: A Large-scale Dataset for Generalized Multi-view Stereo Networks. In CVPR 2020."},{"key":"e_1_3_2_1_56_1","volume-title":"BakedSDF: Meshing Neural SDFs for Real-Time View Synthesis. arXiv preprint arXiv:2302.14859","author":"Yariv Lior","year":"2023","unstructured":"Lior Yariv, Peter Hedman, Christian Reiser, Dor Verbin, Pratul\u00a0P Srinivasan, Richard Szeliski, Jonathan\u00a0T Barron, and Ben Mildenhall. 2023. BakedSDF: Meshing Neural SDFs for Real-Time View Synthesis. arXiv preprint arXiv:2302.14859 (2023)."},{"key":"e_1_3_2_1_57_1","volume-title":"CVPR 2022. 1835","author":"Yuan Yu-Jie","year":"1836","unstructured":"Yu-Jie Yuan, Yang-Tian Sun, Yu-Kun Lai, and et al.2022. NeRF-editing: geometry editing of neural radiance fields. In CVPR 2022. 18353\u201318364."}],"event":{"name":"SA '23: SIGGRAPH Asia 2023","location":"Sydney NSW Australia","acronym":"SA '23","sponsor":["SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques"]},"container-title":["SIGGRAPH Asia 2023 Conference Papers"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3610548.3618190","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3610548.3618190","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T09:30:54Z","timestamp":1755768654000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3610548.3618190"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,10]]},"references-count":57,"alternative-id":["10.1145\/3610548.3618190","10.1145\/3610548"],"URL":"https:\/\/doi.org\/10.1145\/3610548.3618190","relation":{},"subject":[],"published":{"date-parts":[[2023,12,10]]},"assertion":[{"value":"2023-12-11","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}