{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,10]],"date-time":"2026-03-10T01:10:26Z","timestamp":1773105026953,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":65,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,10,28]],"date-time":"2022-10-28T00:00:00Z","timestamp":1666915200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"National Science Foundation Graduate Research Fellowship Program","award":["NSF DGE - 1644869"],"award-info":[{"award-number":["NSF DGE - 1644869"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,10,29]]},"DOI":"10.1145\/3526113.3545621","type":"proceedings-article","created":{"date-parts":[[2022,10,28]],"date-time":"2022-10-28T16:37:41Z","timestamp":1666975061000},"page":"1-17","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":86,"title":["Opal: Multimodal Image Generation for News Illustration"],"prefix":"10.1145","author":[{"given":"Vivian","family":"Liu","sequence":"first","affiliation":[{"name":"Columbia University, United States"}]},{"given":"Han","family":"Qiao","sequence":"additional","affiliation":[{"name":"University of Toronto, Canada and Columbia University, USA"}]},{"given":"Lydia","family":"Chilton","sequence":"additional","affiliation":[{"name":"Computer Science Department, Columbia University, United States"}]}],"member":"320","published-online":{"date-parts":[[2022,10,28]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"[n.d.]. Greater Creative Control for AI image generation. https:\/\/ai.facebook.com\/blog\/greater-creative-control-for-ai-image-generation\/  [n.d.]. Greater Creative Control for AI image generation. https:\/\/ai.facebook.com\/blog\/greater-creative-control-for-ai-image-generation\/"},{"key":"e_1_3_2_1_2_1","volume-title":"HI, USA)","unstructured":"2020. CHI \u201920: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (Honolulu , HI, USA) . Association for Computing Machinery , New York, NY, USA . 2020. CHI \u201920: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (Honolulu, HI, USA). Association for Computing Machinery, New York, NY, USA."},{"key":"e_1_3_2_1_3_1","unstructured":"2022. The Year in Illustration. https:\/\/www.nytimes.com\/interactive\/2022\/01\/05\/multimedia\/year-best-illustration-2021.html  2022. The Year in Illustration. https:\/\/www.nytimes.com\/interactive\/2022\/01\/05\/multimedia\/year-best-illustration-2021.html"},{"key":"e_1_3_2_1_4_1","unstructured":"Adverb. 2021. Advadnoun. https:\/\/twitter.com\/advadnoun  Adverb. 2021. Advadnoun. https:\/\/twitter.com\/advadnoun"},{"key":"e_1_3_2_1_5_1","unstructured":"Gunjan Aggarwal and Devi Parikh. 2020. Neuro-Symbolic Generative Art: A Preliminary Study. arxiv:2007.02171\u00a0[cs.AI]  Gunjan Aggarwal and Devi Parikh. 2020. Neuro-Symbolic Generative Art: A Preliminary Study. arxiv:2007.02171\u00a0[cs.AI]"},{"key":"e_1_3_2_1_6_1","volume-title":"Style follows content: On the microgenesis of art perception. Acta psychologica 128 (06","author":"Augustin M","year":"2008","unstructured":"M Augustin , Helmut Leder , Florian Hutzler , and Claus-Christian Carbon . 2008. Style follows content: On the microgenesis of art perception. Acta psychologica 128 (06 2008 ), 127\u201338. https:\/\/doi.org\/10.1016\/j.actpsy.2007.11.006 10.1016\/j.actpsy.2007.11.006 M Augustin, Helmut Leder, Florian Hutzler, and Claus-Christian Carbon. 2008. Style follows content: On the microgenesis of art perception. Acta psychologica 128 (06 2008), 127\u201338. https:\/\/doi.org\/10.1016\/j.actpsy.2007.11.006"},{"key":"e_1_3_2_1_7_1","unstructured":"Gwern Branwen. 2020. Gpt-3 creative fiction. https:\/\/www.gwern.net\/GPT-3  Gwern Branwen. 2020. Gpt-3 creative fiction. https:\/\/www.gwern.net\/GPT-3"},{"key":"#cr-split#-e_1_3_2_1_8_1.1","unstructured":"Tom\u00a0B. Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell Sandhini Agarwal Ariel Herbert-Voss Gretchen Krueger Tom Henighan Rewon Child Aditya Ramesh Daniel\u00a0M. Ziegler Jeffrey Wu Clemens Winter Christopher Hesse Mark Chen Eric Sigler Mateusz Litwin Scott Gray Benjamin Chess Jack Clark Christopher Berner Sam McCandlish Alec Radford Ilya Sutskever and Dario Amodei. 2020. Language Models are Few-Shot Learners. https:\/\/doi.org\/10.48550\/ARXIV.2005.14165 10.48550\/ARXIV.2005.14165"},{"key":"#cr-split#-e_1_3_2_1_8_1.2","unstructured":"Tom\u00a0B. Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell Sandhini Agarwal Ariel Herbert-Voss Gretchen Krueger Tom Henighan Rewon Child Aditya Ramesh Daniel\u00a0M. Ziegler Jeffrey Wu Clemens Winter Christopher Hesse Mark Chen Eric Sigler Mateusz Litwin Scott Gray Benjamin Chess Jack Clark Christopher Berner Sam McCandlish Alec Radford Ilya Sutskever and Dario Amodei. 2020. Language Models are Few-Shot Learners. https:\/\/doi.org\/10.48550\/ARXIV.2005.14165"},{"key":"e_1_3_2_1_9_1","volume-title":"How Novelists Use Generative Language Models: An Exploratory User Study. In 23rd International Conference on Intelligent User Interfaces. ACM.","author":"Calderwood Alex","year":"2018","unstructured":"Alex Calderwood , Vivian Qiu , Katy\u00a0Ilonka Gero , and Lydia\u00a0 B Chilton . 2018 . How Novelists Use Generative Language Models: An Exploratory User Study. In 23rd International Conference on Intelligent User Interfaces. ACM. Alex Calderwood, Vivian Qiu, Katy\u00a0Ilonka Gero, and Lydia\u00a0B Chilton. 2018. How Novelists Use Generative Language Models: An Exploratory User Study. In 23rd International Conference on Intelligent User Interfaces. ACM."},{"key":"e_1_3_2_1_10_1","volume-title":"MERMAID: Metaphor Generation with Symbolism and Discriminative Decoding. In Proceedings of the 2021 Conference of the North American","author":"Chakrabarty Tuhin","year":"2021","unstructured":"Tuhin Chakrabarty , Xurui Zhang , Smaranda Muresan , and Nanyun Peng . 2021 . MERMAID: Metaphor Generation with Symbolism and Discriminative Decoding. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics , Online , 4250\u20134261. https:\/\/doi.org\/10.18653\/v1\/2021.naacl-main.336 10.18653\/v1 Tuhin Chakrabarty, Xurui Zhang, Smaranda Muresan, and Nanyun Peng. 2021. MERMAID: Metaphor Generation with Symbolism and Discriminative Decoding. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics, Online, 4250\u20134261. https:\/\/doi.org\/10.18653\/v1\/2021.naacl-main.336"},{"key":"e_1_3_2_1_11_1","unstructured":"Angel\u00a0X. Chang Mihail Eric Manolis Savva and Christopher\u00a0D. Manning. [n.d.]. SceneSeer: 3D Scene Design with Natural Language. https:\/\/doi.org\/10.48550\/ARXIV.1703.00050    10.48550\/ARXIV.1703.00050\nAngel\u00a0X. Chang Mihail Eric Manolis Savva and Christopher\u00a0D. Manning. [n.d.]. SceneSeer: 3D Scene Design with Natural Language. https:\/\/doi.org\/10.48550\/ARXIV.1703.00050"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/2501988.2502008"},{"key":"#cr-split#-e_1_3_2_1_13_1.1","unstructured":"Jaemin Cho Abhay Zala and Mohit Bansal. 2022. DALL-Eval: Probing the Reasoning Skills and Social Biases of Text-to-Image Generative Transformers. https:\/\/doi.org\/10.48550\/ARXIV.2202.04053 10.48550\/ARXIV.2202.04053"},{"key":"#cr-split#-e_1_3_2_1_13_1.2","unstructured":"Jaemin Cho Abhay Zala and Mohit Bansal. 2022. DALL-Eval: Probing the Reasoning Skills and Social Biases of Text-to-Image Generative Transformers. https:\/\/doi.org\/10.48550\/ARXIV.2202.04053"},{"key":"#cr-split#-e_1_3_2_1_14_1.1","doi-asserted-by":"crossref","unstructured":"Bob Coyne and Richard Sproat. 2022. WordsEye: an automatic text-to-scene conversion system. https:\/\/doi.org\/10.1145\/383259.383316 10.1145\/383259.383316","DOI":"10.1145\/383259.383316"},{"key":"#cr-split#-e_1_3_2_1_14_1.2","doi-asserted-by":"crossref","unstructured":"Bob Coyne and Richard Sproat. 2022. WordsEye: an automatic text-to-scene conversion system. https:\/\/doi.org\/10.1145\/383259.383316","DOI":"10.1145\/383259.383316"},{"key":"e_1_3_2_1_15_1","unstructured":"Katherine Crowson. 2021. afiaka87\/clip-guided-diffusion: A CLI tool\/python module for generating images from text using guided diffusion and CLIP from OpenAI.https:\/\/github.com\/afiaka87\/clip-guided-diffusion  Katherine Crowson. 2021. afiaka87\/clip-guided-diffusion: A CLI tool\/python module for generating images from text using guided diffusion and CLIP from OpenAI.https:\/\/github.com\/afiaka87\/clip-guided-diffusion"},{"key":"e_1_3_2_1_16_1","unstructured":"Katherine Crowson. 2021. Rivers Have Wings. https:\/\/twitter.com\/RiversHaveWings  Katherine Crowson. 2021. Rivers Have Wings. https:\/\/twitter.com\/RiversHaveWings"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"crossref","unstructured":"Katherine Crowson Stella Biderman Daniel Kornis Dashiell Stander Eric Hallahan Louis Castricato and Edward Raff. 2022. VQGAN-CLIP: Open Domain Image Generation and Editing with Natural Language Guidance. arXiv preprint arXiv:2204.08583(2022).  Katherine Crowson Stella Biderman Daniel Kornis Dashiell Stander Eric Hallahan Louis Castricato and Edward Raff. 2022. VQGAN-CLIP: Open Domain Image Generation and Editing with Natural Language Guidance. arXiv preprint arXiv:2204.08583(2022).","DOI":"10.1007\/978-3-031-19836-6_6"},{"key":"e_1_3_2_1_18_1","volume-title":"Viewing artworks: Contributions of cognitive control and perceptual facilitation to aesthetic experience. Brain and Cognition 70 (06","author":"Cupchik Gerald","year":"2009","unstructured":"Gerald Cupchik , Oshin Vartanian , Adrian Crawley , and David Mikulis . 2009. Viewing artworks: Contributions of cognitive control and perceptual facilitation to aesthetic experience. Brain and Cognition 70 (06 2009 ), 84\u201391. https:\/\/doi.org\/10.1016\/j.bandc.2009.01.003 10.1016\/j.bandc.2009.01.003 Gerald Cupchik, Oshin Vartanian, Adrian Crawley, and David Mikulis. 2009. Viewing artworks: Contributions of cognitive control and perceptual facilitation to aesthetic experience. Brain and Cognition 70 (06 2009), 84\u201391. https:\/\/doi.org\/10.1016\/j.bandc.2009.01.003"},{"key":"#cr-split#-e_1_3_2_1_19_1.1","unstructured":"Boris Dayma Suraj Patil Pedro Cuenca Khalid Saifullah Tanishq Abraham Ph\u00fac Le\u00a0Khac Luke Melas and Ritobrata Ghosh. 2021. DALLE Mini. https:\/\/doi.org\/10.5281\/zenodo.1234 10.5281\/zenodo.1234"},{"key":"#cr-split#-e_1_3_2_1_19_1.2","unstructured":"Boris Dayma Suraj Patil Pedro Cuenca Khalid Saifullah Tanishq Abraham Ph\u00fac Le\u00a0Khac Luke Melas and Ritobrata Ghosh. 2021. DALLE Mini. https:\/\/doi.org\/10.5281\/zenodo.1234"},{"key":"e_1_3_2_1_20_1","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arxiv:1810.04805\u00a0[cs.CL]","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin , Ming-Wei Chang , Kenton Lee , and Kristina Toutanova . 2019 . BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arxiv:1810.04805\u00a0[cs.CL] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arxiv:1810.04805\u00a0[cs.CL]"},{"key":"e_1_3_2_1_21_1","unstructured":"Alaa El-Nouby Shikhar Sharma Hannes Schulz R\u00a0Devon Hjelm Layla El\u00a0Asri Samira\u00a0Ebrahimi Kahou Y. Bengio and Graham Taylor. 2018. Keep Drawing It: Iterative language-based image generation and editing.  Alaa El-Nouby Shikhar Sharma Hannes Schulz R\u00a0Devon Hjelm Layla El\u00a0Asri Samira\u00a0Ebrahimi Kahou Y. Bengio and Graham Taylor. 2018. Keep Drawing It: Iterative language-based image generation and editing."},{"key":"e_1_3_2_1_22_1","unstructured":"Patrick Esser Robin Rombach and Bj\u00f6rn Ommer. [n.d.]. Taming Transformers for High-Resolution Image Synthesis. https:\/\/doi.org\/10.48550\/ARXIV.2012.09841    10.48550\/ARXIV.2012.09841\nPatrick Esser Robin Rombach and Bj\u00f6rn Ommer. [n.d.]. Taming Transformers for High-Resolution Image Synthesis. https:\/\/doi.org\/10.48550\/ARXIV.2012.09841"},{"key":"e_1_3_2_1_23_1","volume-title":"Making Pre-trained Language Models Better Few-shot Learners. arXiv:2012.15723 [cs] (June","author":"Gao Tianyu","year":"2021","unstructured":"Tianyu Gao , Adam Fisch , and Danqi Chen . 2021. Making Pre-trained Language Models Better Few-shot Learners. arXiv:2012.15723 [cs] (June 2021 ). http:\/\/arxiv.org\/abs\/2012.15723 arXiv:2012.15723. Tianyu Gao, Adam Fisch, and Danqi Chen. 2021. Making Pre-trained Language Models Better Few-shot Learners. arXiv:2012.15723 [cs] (June 2021). http:\/\/arxiv.org\/abs\/2012.15723 arXiv:2012.15723."},{"key":"e_1_3_2_1_24_1","unstructured":"Leon\u00a0A. Gatys Alexander\u00a0S. Ecker and Matthias Bethge. 2015. A Neural Algorithm of Artistic Style. arxiv:1508.06576\u00a0[cs.CV]  Leon\u00a0A. Gatys Alexander\u00a0S. Ecker and Matthias Bethge. 2015. A Neural Algorithm of Artistic Style. arxiv:1508.06576\u00a0[cs.CV]"},{"key":"e_1_3_2_1_25_1","unstructured":"Songwei Ge and Devi Parikh. 2021. Visual Conceptual Blending with Large-scale Language and Vision Models. arxiv:2106.14127\u00a0[cs.CL]  Songwei Ge and Devi Parikh. 2021. Visual Conceptual Blending with Large-scale Language and Vision Models. arxiv:2106.14127\u00a0[cs.CL]"},{"key":"e_1_3_2_1_26_1","volume-title":"Mental Models of AI Agents in a Cooperative Game Setting","author":"Gero Katy\u00a0Ilonka","unstructured":"Katy\u00a0Ilonka Gero , Zahra Ashktorab , Casey Dugan , Qian Pan , James Johnson , Werner Geyer , Maria Ruiz , Sarah Miller , David\u00a0 R. Millen , Murray Campbell , Sadhana Kumaravel , and Wei Zhang . 2020. Mental Models of AI Agents in a Cooperative Game Setting . Association for Computing Machinery , New York, NY, USA , 1\u201312. https:\/\/doi.org\/10.1145\/3313831.3376316 10.1145\/3313831.3376316 Katy\u00a0Ilonka Gero, Zahra Ashktorab, Casey Dugan, Qian Pan, James Johnson, Werner Geyer, Maria Ruiz, Sarah Miller, David\u00a0R. Millen, Murray Campbell, Sadhana Kumaravel, and Wei Zhang. 2020. Mental Models of AI Agents in a Cooperative Game Setting. Association for Computing Machinery, New York, NY, USA, 1\u201312. https:\/\/doi.org\/10.1145\/3313831.3376316"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300526"},{"key":"e_1_3_2_1_28_1","volume-title":"Sparks: Inspiration for Science Writing using Language Models. https:\/\/doi.org\/10.48550\/ARXIV.2110.07640","author":"Gero Katy\u00a0Ilonka","year":"2021","unstructured":"Katy\u00a0Ilonka Gero , Vivian Liu , and Lydia\u00a0 B. Chilton . 2021 . Sparks: Inspiration for Science Writing using Language Models. https:\/\/doi.org\/10.48550\/ARXIV.2110.07640 10.48550\/ARXIV.2110.07640 Katy\u00a0Ilonka Gero, Vivian Liu, and Lydia\u00a0B. Chilton. 2021. Sparks: Inspiration for Science Writing using Language Models. https:\/\/doi.org\/10.48550\/ARXIV.2110.07640"},{"key":"e_1_3_2_1_29_1","volume":"202","author":"Jiang Ellen","unstructured":"Ellen Jiang , Kristen Olson , Edwin Toh , Alejandra Molina , Aaron Donsbach , Michael Terry , and Carrie\u00a0 J Cai. 202 2. PromptMaker: Prompt-based Prototyping with Large Language Models. https:\/\/doi.org\/10.1145\/3491101.3503564 10.1145\/3491101.3503564 Ellen Jiang, Kristen Olson, Edwin Toh, Alejandra Molina, Aaron Donsbach, Michael Terry, and Carrie\u00a0J Cai. 2022. PromptMaker: Prompt-based Prototyping with Large Language Models. https:\/\/doi.org\/10.1145\/3491101.3503564","journal-title":"J Cai."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"crossref","unstructured":"Tero Karras Samuli Laine Miika Aittala Janne Hellsten Jaakko Lehtinen and Timo Aila. 2020. Analyzing and Improving the Image Quality of StyleGAN. arxiv:1912.04958\u00a0[cs.CV]  Tero Karras Samuli Laine Miika Aittala Janne Hellsten Jaakko Lehtinen and Timo Aila. 2020. Analyzing and Improving the Image Quality of StyleGAN. arxiv:1912.04958\u00a0[cs.CV]","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"e_1_3_2_1_31_1","volume-title":"Neural correlates of beauty.Journal of neurophysiology 91 4","author":"Kawabata Hideaki","year":"2004","unstructured":"Hideaki Kawabata and Semir Zeki . 2004. Neural correlates of beauty.Journal of neurophysiology 91 4 ( 2004 ), 1699\u2013705. Hideaki Kawabata and Semir Zeki. 2004. Neural correlates of beauty.Journal of neurophysiology 91 4 (2004), 1699\u2013705."},{"key":"e_1_3_2_1_32_1","volume-title":"Prefix-Tuning: Optimizing Continuous Prompts for Generation. arXiv:2101.00190 [cs] (Jan","author":"Li Xiang\u00a0Lisa","year":"2021","unstructured":"Xiang\u00a0Lisa Li and Percy Liang . 2021. Prefix-Tuning: Optimizing Continuous Prompts for Generation. arXiv:2101.00190 [cs] (Jan . 2021 ). http:\/\/arxiv.org\/abs\/2101.00190 arXiv:2101.00190. Xiang\u00a0Lisa Li and Percy Liang. 2021. Prefix-Tuning: Optimizing Continuous Prompts for Generation. arXiv:2101.00190 [cs] (Jan. 2021). http:\/\/arxiv.org\/abs\/2101.00190 arXiv:2101.00190."},{"key":"e_1_3_2_1_33_1","unstructured":"Vivian Liu and Lydia Chilton. [n.d.]. Neurosymbolic generation of 3D animal shapes through... - ceur-ws.org. http:\/\/ceur-ws.org\/Vol-2903\/IUI21WS-HAIGEN-8.pdf  Vivian Liu and Lydia Chilton. [n.d.]. Neurosymbolic generation of 3D animal shapes through... - ceur-ws.org. http:\/\/ceur-ws.org\/Vol-2903\/IUI21WS-HAIGEN-8.pdf"},{"key":"e_1_3_2_1_34_1","unstructured":"Vivian Liu and Lydia\u00a0B. Chilton. 2021. Design Guidelines for Prompt Engineering Text-to-Image Generative Models. arxiv:2109.06977\u00a0[cs.HC]  Vivian Liu and Lydia\u00a0B. Chilton. 2021. Design Guidelines for Prompt Engineering Text-to-Image Generative Models. arxiv:2109.06977\u00a0[cs.HC]"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3173574.3173943"},{"key":"e_1_3_2_1_36_1","unstructured":"Ryan Murdock. [n.d.]. lucidrains\/big-sleep: A simple command line tool for text to image generation using OpenAI\u2019s CLIP and a BigGAN. Technique was originally created by https:\/\/twitter.com\/advadnoun. https:\/\/github.com\/lucidrains\/big-sleep  Ryan Murdock. [n.d.]. lucidrains\/big-sleep: A simple command line tool for text to image generation using OpenAI\u2019s CLIP and a BigGAN. Technique was originally created by https:\/\/twitter.com\/advadnoun. https:\/\/github.com\/lucidrains\/big-sleep"},{"key":"e_1_3_2_1_37_1","unstructured":"nerdyroden. 2022. nerdyrodent\/VQGAN-CLIP. https:\/\/github.com\/nerdyrodent\/VQGAN-CLIP  nerdyroden. 2022. nerdyrodent\/VQGAN-CLIP. https:\/\/github.com\/nerdyrodent\/VQGAN-CLIP"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00244"},{"key":"e_1_3_2_1_39_1","volume-title":"Advances in Neural Information Processing Systems, H.\u00a0Wallach, H.\u00a0Larochelle, A.\u00a0Beygelzimer, F.\u00a0d'Alch\u00e9-Buc, E.\u00a0Fox, and R.\u00a0Garnett (Eds.). Vol.\u00a032. Curran Associates","author":"Qiao Tingting","year":"2019","unstructured":"Tingting Qiao , Jing Zhang , Duanqing Xu , and Dacheng Tao . 2019. Learn , Imagine and Create: Text-to-Image Generation from Prior Knowledge . In Advances in Neural Information Processing Systems, H.\u00a0Wallach, H.\u00a0Larochelle, A.\u00a0Beygelzimer, F.\u00a0d'Alch\u00e9-Buc, E.\u00a0Fox, and R.\u00a0Garnett (Eds.). Vol.\u00a032. Curran Associates , Inc .https:\/\/proceedings.neurips.cc\/paper\/ 2019 \/file\/d18f655c3fce66ca401d5f38b48c89af-Paper.pdf Tingting Qiao, Jing Zhang, Duanqing Xu, and Dacheng Tao. 2019. Learn, Imagine and Create: Text-to-Image Generation from Prior Knowledge. In Advances in Neural Information Processing Systems, H.\u00a0Wallach, H.\u00a0Larochelle, A.\u00a0Beygelzimer, F.\u00a0d'Alch\u00e9-Buc, E.\u00a0Fox, and R.\u00a0Garnett (Eds.). Vol.\u00a032. Curran Associates, Inc.https:\/\/proceedings.neurips.cc\/paper\/2019\/file\/d18f655c3fce66ca401d5f38b48c89af-Paper.pdf"},{"key":"#cr-split#-e_1_3_2_1_40_1.1","doi-asserted-by":"crossref","unstructured":"Tingting Qiao Jing Zhang Duanqing Xu and Dacheng Tao. 2019. MirrorGAN: Learning Text-to-image Generation by Redescription. https:\/\/doi.org\/10.48550\/ARXIV.1903.05854 10.48550\/ARXIV.1903.05854","DOI":"10.1109\/CVPR.2019.00160"},{"key":"#cr-split#-e_1_3_2_1_40_1.2","doi-asserted-by":"crossref","unstructured":"Tingting Qiao Jing Zhang Duanqing Xu and Dacheng Tao. 2019. MirrorGAN: Learning Text-to-image Generation by Redescription. https:\/\/doi.org\/10.48550\/ARXIV.1903.05854","DOI":"10.1109\/CVPR.2019.00160"},{"key":"e_1_3_2_1_41_1","unstructured":"Alec Radford Jong\u00a0Wook Kim Chris Hallacy Aditya Ramesh Gabriel Goh Sandhini Agarwal Girish Sastry Amanda Askell Pamela Mishkin Jack Clark Gretchen Krueger and Ilya Sutskever. 2021. Learning Transferable Visual Models From Natural Language Supervision. arxiv:2103.00020\u00a0[cs.CV]  Alec Radford Jong\u00a0Wook Kim Chris Hallacy Aditya Ramesh Gabriel Goh Sandhini Agarwal Girish Sastry Amanda Askell Pamela Mishkin Jack Clark Gretchen Krueger and Ilya Sutskever. 2021. Learning Transferable Visual Models From Natural Language Supervision. arxiv:2103.00020\u00a0[cs.CV]"},{"key":"#cr-split#-e_1_3_2_1_42_1.1","unstructured":"Aditya Ramesh Prafulla Dhariwal Alex Nichol Casey Chu and Mark Chen. 2022. Hierarchical Text-Conditional Image Generation with CLIP Latents. https:\/\/doi.org\/10.48550\/ARXIV.2204.06125 10.48550\/ARXIV.2204.06125"},{"key":"#cr-split#-e_1_3_2_1_42_1.2","unstructured":"Aditya Ramesh Prafulla Dhariwal Alex Nichol Casey Chu and Mark Chen. 2022. Hierarchical Text-Conditional Image Generation with CLIP Latents. https:\/\/doi.org\/10.48550\/ARXIV.2204.06125"},{"key":"e_1_3_2_1_43_1","unstructured":"[\n  43\n  ]  reddit.com.2021. https:\/\/www.reddit.com\/r\/bigsleep\/  [43] reddit.com.2021. https:\/\/www.reddit.com\/r\/bigsleep\/"},{"key":"#cr-split#-e_1_3_2_1_44_1.1","doi-asserted-by":"crossref","unstructured":"Nils Reimers and Iryna Gurevych. 2019. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. https:\/\/doi.org\/10.48550\/ARXIV.1908.10084 10.48550\/ARXIV.1908.10084","DOI":"10.18653\/v1\/D19-1410"},{"key":"#cr-split#-e_1_3_2_1_44_1.2","doi-asserted-by":"crossref","unstructured":"Nils Reimers and Iryna Gurevych. 2019. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. https:\/\/doi.org\/10.48550\/ARXIV.1908.10084","DOI":"10.18653\/v1\/D19-1410"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"crossref","unstructured":"Laria Reynolds and Kyle McDonell. 2021. Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm. arxiv:2102.07350\u00a0[cs.CL]  Laria Reynolds and Kyle McDonell. 2021. Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm. arxiv:2102.07350\u00a0[cs.CL]","DOI":"10.1145\/3411763.3451760"},{"key":"e_1_3_2_1_46_1","volume-title":"Burcu\u00a0Karagol Ayan, S.\u00a0Sara Mahdavi, Rapha\u00a0Gontijo Lopes, Tim Salimans, Jonathan Ho, David\u00a0J Fleet, and Mohammad Norouzi.","author":"Saharia Chitwan","year":"2022","unstructured":"Chitwan Saharia , William Chan , Saurabh Saxena , Lala Li , Jay Whang , Emily Denton , Seyed Kamyar\u00a0Seyed Ghasemipour , Burcu\u00a0Karagol Ayan, S.\u00a0Sara Mahdavi, Rapha\u00a0Gontijo Lopes, Tim Salimans, Jonathan Ho, David\u00a0J Fleet, and Mohammad Norouzi. 2022 . Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding . https:\/\/doi.org\/10.48550\/ARXIV.2205.11487 10.48550\/ARXIV.2205.11487 Chitwan Saharia, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily Denton, Seyed Kamyar\u00a0Seyed Ghasemipour, Burcu\u00a0Karagol Ayan, S.\u00a0Sara Mahdavi, Rapha\u00a0Gontijo Lopes, Tim Salimans, Jonathan Ho, David\u00a0J Fleet, and Mohammad Norouzi. 2022. Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding. https:\/\/doi.org\/10.48550\/ARXIV.2205.11487"},{"key":"#cr-split#-e_1_3_2_1_47_1.1","unstructured":"Shikhar Sharma Dendi Suhubdy Vincent Michalski Samira\u00a0Ebrahimi Kahou and Yoshua Bengio. 2018. ChatPainter: Improving Text to Image Generation using Dialogue. https:\/\/doi.org\/10.48550\/ARXIV.1802.08216 10.48550\/ARXIV.1802.08216"},{"key":"#cr-split#-e_1_3_2_1_47_1.2","unstructured":"Shikhar Sharma Dendi Suhubdy Vincent Michalski Samira\u00a0Ebrahimi Kahou and Yoshua Bengio. 2018. ChatPainter: Improving Text to Image Generation using Dialogue. https:\/\/doi.org\/10.48550\/ARXIV.1802.08216"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/3379337.3415866"},{"key":"e_1_3_2_1_49_1","volume-title":"Where to Hide a Stolen Elephant: Leaps in Creative Writing with Multimodal Machine Intelligence. ACM Trans. Comput.-Hum. Interact. (jan","author":"Singh Nikhil","year":"2022","unstructured":"Nikhil Singh , Guillermo Bernal , Daria Savchenko , and Elena\u00a0 L. Glassman . 2022. Where to Hide a Stolen Elephant: Leaps in Creative Writing with Multimodal Machine Intelligence. ACM Trans. Comput.-Hum. Interact. (jan 2022 ). https:\/\/doi.org\/10.1145\/3511599 Just Accepted . 10.1145\/3511599 Nikhil Singh, Guillermo Bernal, Daria Savchenko, and Elena\u00a0L. Glassman. 2022. Where to Hide a Stolen Elephant: Leaps in Creative Writing with Multimodal Machine Intelligence. ACM Trans. Comput.-Hum. Interact. (jan 2022). https:\/\/doi.org\/10.1145\/3511599 Just Accepted."},{"key":"e_1_3_2_1_50_1","volume":"202","author":"Wu Tongshuang","unstructured":"Tongshuang Wu , Ellen Jiang , Aaron Donsbach , Jeff Gray , Alejandra Molina , Michael Terry , and Carrie\u00a0 J Cai. 202 2. PromptChainer: Chaining Large Language Model Prompts through Visual Programming. https:\/\/doi.org\/10.48550\/ARXIV.2203.06566 10.48550\/ARXIV.2203.06566 Tongshuang Wu, Ellen Jiang, Aaron Donsbach, Jeff Gray, Alejandra Molina, Michael Terry, and Carrie\u00a0J Cai. 2022. PromptChainer: Chaining Large Language Model Prompts through Visual Programming. https:\/\/doi.org\/10.48550\/ARXIV.2203.06566","journal-title":"J Cai."},{"key":"e_1_3_2_1_51_1","volume":"202","author":"Wu Tongshuang","unstructured":"Tongshuang Wu , Michael Terry , and Carrie\u00a0 J Cai. 202 2. AI Chains: Transparent and Controllable Human-AI Interaction by Chaining Large Language Model Prompts. https:\/\/doi.org\/10.1145\/3491102.3517582 10.1145\/3491102.3517582 Tongshuang Wu, Michael Terry, and Carrie\u00a0J Cai. 2022. AI Chains: Transparent and Controllable Human-AI Interaction by Chaining Large Language Model Prompts. https:\/\/doi.org\/10.1145\/3491102.3517582","journal-title":"J Cai."},{"key":"#cr-split#-e_1_3_2_1_52_1.1","unstructured":"Weihao Xia Yujiu Yang Jing-Hao Xue and Baoyuan Wu. 2020. TediGAN: Text-Guided Diverse Face Image Generation and Manipulation. https:\/\/doi.org\/10.48550\/ARXIV.2012.03308 10.48550\/ARXIV.2012.03308"},{"key":"#cr-split#-e_1_3_2_1_52_1.2","unstructured":"Weihao Xia Yujiu Yang Jing-Hao Xue and Baoyuan Wu. 2020. TediGAN: Text-Guided Diverse Face Image Generation and Manipulation. https:\/\/doi.org\/10.48550\/ARXIV.2012.03308"},{"key":"#cr-split#-e_1_3_2_1_53_1.1","unstructured":"Tao Xu Pengchuan Zhang Qiuyuan Huang Han Zhang Zhe Gan Xiaolei Huang and Xiaodong He. 2017. AttnGAN: Fine-Grained Text to Image Generation with Attentional Generative Adversarial Networks. https:\/\/doi.org\/10.48550\/ARXIV.1711.10485 10.48550\/ARXIV.1711.10485"},{"key":"#cr-split#-e_1_3_2_1_53_1.2","doi-asserted-by":"crossref","unstructured":"Tao Xu Pengchuan Zhang Qiuyuan Huang Han Zhang Zhe Gan Xiaolei Huang and Xiaodong He. 2017. AttnGAN: Fine-Grained Text to Image Generation with Attentional Generative Adversarial Networks. https:\/\/doi.org\/10.48550\/ARXIV.1711.10485","DOI":"10.1109\/CVPR.2018.00143"},{"key":"#cr-split#-e_1_3_2_1_54_1.1","unstructured":"Jiahui Yu Yuanzhong Xu Jing\u00a0Yu Koh Thang Luong Gunjan Baid Zirui Wang Vijay Vasudevan Alexander Ku Yinfei Yang Burcu\u00a0Karagol Ayan Ben Hutchinson Wei Han Zarana Parekh Xin Li Han Zhang Jason Baldridge and Yonghui Wu. 2022. Scaling Autoregressive Models for Content-Rich Text-to-Image Generation. https:\/\/doi.org\/10.48550\/ARXIV.2206.10789 10.48550\/ARXIV.2206.10789"},{"key":"#cr-split#-e_1_3_2_1_54_1.2","unstructured":"Jiahui Yu Yuanzhong Xu Jing\u00a0Yu Koh Thang Luong Gunjan Baid Zirui Wang Vijay Vasudevan Alexander Ku Yinfei Yang Burcu\u00a0Karagol Ayan Ben Hutchinson Wei Han Zarana Parekh Xin Li Han Zhang Jason Baldridge and Yonghui Wu. 2022. Scaling Autoregressive Models for Content-Rich Text-to-Image Generation. https:\/\/doi.org\/10.48550\/ARXIV.2206.10789"}],"event":{"name":"UIST '22: The 35th Annual ACM Symposium on User Interface Software and Technology","location":"Bend OR USA","acronym":"UIST '22","sponsor":["SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques","SIGCHI ACM Special Interest Group on Computer-Human Interaction"]},"container-title":["Proceedings of the 35th Annual ACM Symposium on User Interface Software and Technology"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3526113.3545621","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3526113.3545621","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:00:23Z","timestamp":1750186823000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3526113.3545621"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,10,28]]},"references-count":65,"alternative-id":["10.1145\/3526113.3545621","10.1145\/3526113"],"URL":"https:\/\/doi.org\/10.1145\/3526113.3545621","relation":{},"subject":[],"published":{"date-parts":[[2022,10,28]]},"assertion":[{"value":"2022-10-28","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}