{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T02:47:36Z","timestamp":1774925256888,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":30,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,7,23]],"date-time":"2023-07-23T00:00:00Z","timestamp":1690070400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,7,23]]},"DOI":"10.1145\/3588432.3591506","type":"proceedings-article","created":{"date-parts":[[2023,7,19]],"date-time":"2023-07-19T13:34:52Z","timestamp":1689773692000},"page":"1-11","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":109,"title":["Key-Locked Rank One Editing for Text-to-Image Personalization"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0006-8042-0428","authenticated-orcid":false,"given":"Yoad","family":"Tewel","sequence":"first","affiliation":[{"name":"NVIDIA Research, Israel"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4875-965X","authenticated-orcid":false,"given":"Rinon","family":"Gal","sequence":"additional","affiliation":[{"name":"NVIDIA Research, Israel"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9164-5303","authenticated-orcid":false,"given":"Gal","family":"Chechik","sequence":"additional","affiliation":[{"name":"NVIDIA Research, Israel"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3817-3698","authenticated-orcid":false,"given":"Yuval","family":"Atzmon","sequence":"additional","affiliation":[{"name":"NVIDIA Research, Israel"}]}],"member":"320","published-online":{"date-parts":[[2023,7,23]]},"reference":[{"key":"e_1_3_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00453"},{"key":"e_1_3_2_2_2_1","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 18208\u201318218","author":"Avrahami Omri","year":"2022","unstructured":"Omri Avrahami , Dani Lischinski , and Ohad Fried . 2022 . Blended diffusion for text-driven editing of natural images . In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 18208\u201318218 . Omri Avrahami, Dani Lischinski, and Ohad Fried. 2022. Blended diffusion for text-driven editing of natural images. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 18208\u201318218."},{"key":"e_1_3_2_2_3_1","volume-title":"eDiff-I: Text-to-Image Diffusion Models with Ensemble of Expert Denoisers. arXiv preprint arXiv:2211.01324","author":"Balaji Yogesh","year":"2022","unstructured":"Yogesh Balaji , Seungjun Nah , Xun Huang , Arash Vahdat , Jiaming Song , Karsten Kreis , Miika Aittala , Timo Aila , Samuli Laine , Bryan Catanzaro , Tero Karras , and Ming-Yu Liu . 2022. eDiff-I: Text-to-Image Diffusion Models with Ensemble of Expert Denoisers. arXiv preprint arXiv:2211.01324 ( 2022 ). Yogesh Balaji, Seungjun Nah, Xun Huang, Arash Vahdat, Jiaming Song, Karsten Kreis, Miika Aittala, Timo Aila, Samuli Laine, Bryan Catanzaro, Tero Karras, and Ming-Yu Liu. 2022. eDiff-I: Text-to-Image Diffusion Models with Ensemble of Expert Denoisers. arXiv preprint arXiv:2211.01324 (2022)."},{"key":"e_1_3_2_2_4_1","volume-title":"Proceedings of the European Conference on Computer Vision (ECCV).","author":"Bau David","year":"2020","unstructured":"David Bau , Steven Liu , Tongzhou Wang , Jun-Yan Zhu , and Antonio Torralba . 2020 . Rewriting a Deep Generative Model . In Proceedings of the European Conference on Computer Vision (ECCV). David Bau, Steven Liu, Tongzhou Wang, Jun-Yan Zhu, and Antonio Torralba. 2020. Rewriting a Deep Generative Model. In Proceedings of the European Conference on Computer Vision (ECCV)."},{"key":"e_1_3_2_2_5_1","volume-title":"InstructPix2Pix: Learning to Follow Image Editing Instructions. ArXiv abs\/2211.09800","author":"Brooks Tim","year":"2022","unstructured":"Tim Brooks , Aleksander Holynski , and Alexei\u00a0 A. Efros . 2022. InstructPix2Pix: Learning to Follow Image Editing Instructions. ArXiv abs\/2211.09800 ( 2022 ). Tim Brooks, Aleksander Holynski, and Alexei\u00a0A. Efros. 2022. InstructPix2Pix: Learning to Follow Image Editing Instructions. ArXiv abs\/2211.09800 (2022)."},{"key":"e_1_3_2_2_6_1","volume-title":"European Conference on Computer Vision (ECCV).","author":"Cohen Niv","year":"2022","unstructured":"Niv Cohen , Rinon Gal , Eli\u00a0 A. Meirom , Gal Chechik , and Yuval Atzmon . 2022 . \" This is my unicorn, Fluffy\": Personalizing frozen vision-language representations . In European Conference on Computer Vision (ECCV). Niv Cohen, Rinon Gal, Eli\u00a0A. Meirom, Gal Chechik, and Yuval Atzmon. 2022. \"This is my unicorn, Fluffy\": Personalizing frozen vision-language representations. In European Conference on Computer Vision (ECCV)."},{"key":"e_1_3_2_2_7_1","volume-title":"Multiresolution Textual Inversion. ArXiv abs\/2211.17115","author":"Daras Giannis","year":"2022","unstructured":"Giannis Daras and Alexandros\u00a0 G. Dimakis . 2022. Multiresolution Textual Inversion. ArXiv abs\/2211.17115 ( 2022 ). Giannis Daras and Alexandros\u00a0G. Dimakis. 2022. Multiresolution Textual Inversion. ArXiv abs\/2211.17115 (2022)."},{"key":"e_1_3_2_2_8_1","unstructured":"darkstorm2150. 2022. Protogen-v3.4. https:\/\/huggingface.co\/darkstorm2150\/Protogen_x3.4_Official_Release\/tree\/main darkstorm2150. 2022. Protogen-v3.4. https:\/\/huggingface.co\/darkstorm2150\/Protogen_x3.4_Official_Release\/tree\/main"},{"key":"e_1_3_2_2_9_1","unstructured":"Envvi. 2022. InkPunk-v2. https:\/\/huggingface.co\/Envvi\/Inkpunk-Diffusion Envvi. 2022. InkPunk-v2. https:\/\/huggingface.co\/Envvi\/Inkpunk-Diffusion"},{"key":"#cr-split#-e_1_3_2_2_10_1.1","unstructured":"Rinon Gal Yuval Alaluf Yuval Atzmon Or Patashnik Amit\u00a0H. Bermano Gal Chechik and Daniel Cohen-Or. 2022. An Image is Worth One Word: Personalizing Text-to-Image Generation using Textual Inversion. https:\/\/doi.org\/10.48550\/ARXIV.2208.01618 10.48550\/ARXIV.2208.01618"},{"key":"#cr-split#-e_1_3_2_2_10_1.2","unstructured":"Rinon Gal Yuval Alaluf Yuval Atzmon Or Patashnik Amit\u00a0H. Bermano Gal Chechik and Daniel Cohen-Or. 2022. An Image is Worth One Word: Personalizing Text-to-Image Generation using Textual Inversion. https:\/\/doi.org\/10.48550\/ARXIV.2208.01618"},{"key":"e_1_3_2_2_11_1","volume-title":"Transformer Feed-Forward Layers Build Predictions by Promoting Concepts in the Vocabulary Space. arXiv preprint arXiv:2203.14680","author":"Geva Mor","year":"2022","unstructured":"Mor Geva , Avi Caciularu , Kevin\u00a0Ro Wang , and Yoav Goldberg . 2022. Transformer Feed-Forward Layers Build Predictions by Promoting Concepts in the Vocabulary Space. arXiv preprint arXiv:2203.14680 ( 2022 ). Mor Geva, Avi Caciularu, Kevin\u00a0Ro Wang, and Yoav Goldberg. 2022. Transformer Feed-Forward Layers Build Predictions by Promoting Concepts in the Vocabulary Space. arXiv preprint arXiv:2203.14680 (2022)."},{"key":"e_1_3_2_2_12_1","volume-title":"Transformer feed-forward layers are key-value memories. arXiv preprint arXiv:2012.14913","author":"Geva Mor","year":"2020","unstructured":"Mor Geva , Roei Schuster , Jonathan Berant , and Omer Levy . 2020. Transformer feed-forward layers are key-value memories. arXiv preprint arXiv:2012.14913 ( 2020 ). Mor Geva, Roei Schuster, Jonathan Berant, and Omer Levy. 2020. Transformer feed-forward layers are key-value memories. arXiv preprint arXiv:2012.14913 (2020)."},{"key":"e_1_3_2_2_13_1","unstructured":"Amir Hertz Ron Mokady Jay Tenenbaum Kfir Aberman Yael Pritch and Daniel Cohen-Or. 2022. Prompt-to-prompt image editing with cross attention control. (2022). Amir Hertz Ron Mokady Jay Tenenbaum Kfir Aberman Yael Pritch and Daniel Cohen-Or. 2022. Prompt-to-prompt image editing with cross attention control. (2022)."},{"key":"e_1_3_2_2_14_1","volume-title":"Imagic: Text-Based Real Image Editing with Diffusion Models. arXiv preprint arXiv:2210.09276","author":"Kawar Bahjat","year":"2022","unstructured":"Bahjat Kawar , Shiran Zada , Oran Lang , Omer Tov , Huiwen Chang , Tali Dekel , Inbar Mosseri , and Michal Irani . 2022 . Imagic: Text-Based Real Image Editing with Diffusion Models. arXiv preprint arXiv:2210.09276 (2022). Bahjat Kawar, Shiran Zada, Oran Lang, Omer Tov, Huiwen Chang, Tali Dekel, Inbar Mosseri, and Michal Irani. 2022. Imagic: Text-Based Real Image Editing with Diffusion Models. arXiv preprint arXiv:2210.09276 (2022)."},{"key":"e_1_3_2_2_15_1","volume-title":"Multi-Concept Customization of Text-to-Image Diffusion. arXiv","author":"Kumari Nupur","year":"2022","unstructured":"Nupur Kumari , Bingliang Zhang , Richard Zhang , Eli Shechtman , and Jun-Yan Zhu . 2022. Multi-Concept Customization of Text-to-Image Diffusion. arXiv ( 2022 ). Nupur Kumari, Bingliang Zhang, Richard Zhang, Eli Shechtman, and Jun-Yan Zhu. 2022. Multi-Concept Customization of Text-to-Image Diffusion. arXiv (2022)."},{"key":"e_1_3_2_2_16_1","volume-title":"Prompt-Based Multi-Modal Image Segmentation. arXiv preprint arXiv:2112.10003","author":"L\u00fcddecke Timo","year":"2021","unstructured":"Timo L\u00fcddecke and Alexander\u00a0 S Ecker . 2021. Prompt-Based Multi-Modal Image Segmentation. arXiv preprint arXiv:2112.10003 ( 2021 ). Timo L\u00fcddecke and Alexander\u00a0S Ecker. 2021. Prompt-Based Multi-Modal Image Segmentation. arXiv preprint arXiv:2112.10003 (2021)."},{"key":"e_1_3_2_2_17_1","volume-title":"International Conference on Learning Representations.","author":"Meng Chenlin","year":"2021","unstructured":"Chenlin Meng , Yutong He , Yang Song , Jiaming Song , Jiajun Wu , Jun-Yan Zhu , and Stefano Ermon . 2021 . Sdedit: Guided image synthesis and editing with stochastic differential equations . In International Conference on Learning Representations. Chenlin Meng, Yutong He, Yang Song, Jiaming Song, Jiajun Wu, Jun-Yan Zhu, and Stefano Ermon. 2021. Sdedit: Guided image synthesis and editing with stochastic differential equations. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_18_1","volume-title":"Locating and Editing Factual Associations in GPT. Advances in Neural Information Processing Systems 36","author":"Meng Kevin","year":"2022","unstructured":"Kevin Meng , David Bau , Alex Andonian , and Yonatan Belinkov . 2022. Locating and Editing Factual Associations in GPT. Advances in Neural Information Processing Systems 36 ( 2022 ). Kevin Meng, David Bau, Alex Andonian, and Yonatan Belinkov. 2022. Locating and Editing Factual Associations in GPT. Advances in Neural Information Processing Systems 36 (2022)."},{"key":"e_1_3_2_2_19_1","volume-title":"Glide: Towards photorealistic image generation and editing with text-guided diffusion models. arXiv preprint arXiv:2112.10741","author":"Nichol Alex","year":"2021","unstructured":"Alex Nichol , Prafulla Dhariwal , Aditya Ramesh , Pranav Shyam , Pamela Mishkin , Bob McGrew , Ilya Sutskever , and Mark Chen . 2021 . Glide: Towards photorealistic image generation and editing with text-guided diffusion models. arXiv preprint arXiv:2112.10741 (2021). Alex Nichol, Prafulla Dhariwal, Aditya Ramesh, Pranav Shyam, Pamela Mishkin, Bob McGrew, Ilya Sutskever, and Mark Chen. 2021. Glide: Towards photorealistic image generation and editing with text-guided diffusion models. arXiv preprint arXiv:2112.10741 (2021)."},{"key":"e_1_3_2_2_20_1","volume-title":"International Conference on Machine Learning. PMLR, 8748\u20138763","author":"Radford Alec","year":"2021","unstructured":"Alec Radford , Jong\u00a0Wook Kim , Chris Hallacy , Aditya Ramesh , Gabriel Goh , Sandhini Agarwal , Girish Sastry , Amanda Askell , Pamela Mishkin , Jack Clark , 2021 . Learning transferable visual models from natural language supervision . In International Conference on Machine Learning. PMLR, 8748\u20138763 . Alec Radford, Jong\u00a0Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, 2021. Learning transferable visual models from natural language supervision. In International Conference on Machine Learning. PMLR, 8748\u20138763."},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.5555\/3455716.3455856"},{"key":"e_1_3_2_2_22_1","volume-title":"Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125","author":"Ramesh Aditya","year":"2022","unstructured":"Aditya Ramesh , Prafulla Dhariwal , Alex Nichol , Casey Chu , and Mark Chen . 2022. Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125 ( 2022 ). Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. 2022. Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125 (2022)."},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"crossref","unstructured":"Robin Rombach Andreas Blattmann Dominik Lorenz Patrick Esser and Bj\u00f6rn Ommer. 2021. High-Resolution Image Synthesis with Latent Diffusion Models. arxiv:2112.10752\u00a0[cs.CV] Robin Rombach Andreas Blattmann Dominik Lorenz Patrick Esser and Bj\u00f6rn Ommer. 2021. High-Resolution Image Synthesis with Latent Diffusion Models. arxiv:2112.10752\u00a0[cs.CV]","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"e_1_3_2_2_24_1","doi-asserted-by":"crossref","unstructured":"Nataniel Ruiz Yuanzhen Li Varun Jampani Yael Pritch Michael Rubinstein and Kfir Aberman. 2022. DreamBooth: Fine Tuning Text-to-image Diffusion Models for Subject-Driven Generation. (2022). Nataniel Ruiz Yuanzhen Li Varun Jampani Yael Pritch Michael Rubinstein and Kfir Aberman. 2022. DreamBooth: Fine Tuning Text-to-image Diffusion Models for Subject-Driven Generation. (2022).","DOI":"10.1109\/CVPR52729.2023.02155"},{"key":"e_1_3_2_2_25_1","volume-title":"Burcu\u00a0Karagol Ayan, S\u00a0Sara Mahdavi","author":"Saharia Chitwan","year":"2022","unstructured":"Chitwan Saharia , William Chan , Saurabh Saxena , Lala Li , Jay Whang , Emily Denton , Seyed Kamyar\u00a0Seyed Ghasemipour , Burcu\u00a0Karagol Ayan, S\u00a0Sara Mahdavi , Rapha\u00a0Gontijo Lopes, 2022 . Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding . arXiv preprint arXiv:2205.11487 (2022). Chitwan Saharia, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily Denton, Seyed Kamyar\u00a0Seyed Ghasemipour, Burcu\u00a0Karagol Ayan, S\u00a0Sara Mahdavi, Rapha\u00a0Gontijo Lopes, 2022. Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding. arXiv preprint arXiv:2205.11487 (2022)."},{"key":"e_1_3_2_2_26_1","volume-title":"2021 IEEE Winter Conference on Applications of Computer Vision (WACV)","author":"Samuel Dvir","year":"2020","unstructured":"Dvir Samuel , Yuval Atzmon , and Gal Chechik . 2020 . From generalized zero-shot learning to long-tail with class descriptors . 2021 IEEE Winter Conference on Applications of Computer Vision (WACV) (2020), 286\u2013295. Dvir Samuel, Yuval Atzmon, and Gal Chechik. 2020. From generalized zero-shot learning to long-tail with class descriptors. 2021 IEEE Winter Conference on Applications of Computer Vision (WACV) (2020), 286\u2013295."},{"key":"e_1_3_2_2_27_1","volume-title":"Uncovering the Disentanglement Capability in Text-to-Image Diffusion Models. ArXiv abs\/2212.08698","author":"Wu Qiucheng","year":"2022","unstructured":"Qiucheng Wu , Yujian Liu , Handong Zhao , Ajinkya Kale , Trung\u00a0 M. Bui , Tong Yu , Zhe Lin , Yang Zhang , and Shiyu Chang . 2022. Uncovering the Disentanglement Capability in Text-to-Image Diffusion Models. ArXiv abs\/2212.08698 ( 2022 ). Qiucheng Wu, Yujian Liu, Handong Zhao, Ajinkya Kale, Trung\u00a0M. Bui, Tong Yu, Zhe Lin, Yang Zhang, and Shiyu Chang. 2022. Uncovering the Disentanglement Capability in Text-to-Image Diffusion Models. ArXiv abs\/2212.08698 (2022)."},{"key":"e_1_3_2_2_28_1","volume-title":"Paint by Example: Exemplar-based Image Editing with Diffusion Models. ArXiv abs\/2211.13227","author":"Yang Binxin","year":"2022","unstructured":"Binxin Yang , Shuyang Gu , Bo Zhang , Ting Zhang , Xuejin Chen , Xiaoyan Sun , Dong Chen , and Fang Wen . 2022. Paint by Example: Exemplar-based Image Editing with Diffusion Models. ArXiv abs\/2211.13227 ( 2022 ). Binxin Yang, Shuyang Gu, Bo Zhang, Ting Zhang, Xuejin Chen, Xiaoyan Sun, Dong Chen, and Fang Wen. 2022. Paint by Example: Exemplar-based Image Editing with Diffusion Models. ArXiv abs\/2211.13227 (2022)."},{"key":"e_1_3_2_2_29_1","volume-title":"SINE: SINgle Image Editing with Text-to-Image Diffusion Models. ArXiv abs\/2212.04489","author":"Zhang Zhixing","year":"2022","unstructured":"Zhixing Zhang , Ligong Han , Arna Ghosh , Dimitris\u00a0 N. Metaxas , and Jian Ren . 2022 . SINE: SINgle Image Editing with Text-to-Image Diffusion Models. ArXiv abs\/2212.04489 (2022). Zhixing Zhang, Ligong Han, Arna Ghosh, Dimitris\u00a0N. Metaxas, and Jian Ren. 2022. SINE: SINgle Image Editing with Text-to-Image Diffusion Models. ArXiv abs\/2212.04489 (2022)."}],"event":{"name":"SIGGRAPH '23: Special Interest Group on Computer Graphics and Interactive Techniques Conference","location":"Los Angeles CA USA","acronym":"SIGGRAPH '23","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction"]},"container-title":["Special Interest Group on Computer Graphics and Interactive Techniques Conference Conference Proceedings"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3588432.3591506","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T17:51:29Z","timestamp":1750182689000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3588432.3591506"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,7,23]]},"references-count":30,"alternative-id":["10.1145\/3588432.3591506","10.1145\/3588432"],"URL":"https:\/\/doi.org\/10.1145\/3588432.3591506","relation":{},"subject":[],"published":{"date-parts":[[2023,7,23]]},"assertion":[{"value":"2023-07-23","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}