{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T18:59:20Z","timestamp":1776106760246,"version":"3.50.1"},"reference-count":109,"publisher":"Association for Computing Machinery (ACM)","issue":"5","license":[{"start":{"date-parts":[[2024,10,31]],"date-time":"2024-10-31T00:00:00Z","timestamp":1730332800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-sa\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Comput.-Hum. Interact."],"published-print":{"date-parts":[[2024,10,31]]},"abstract":"<jats:p>\n            Providing asynchronous feedback is a critical step in the 3D design workflow. A common approach to providing feedback is to pair textual comments with companion reference images, which helps illustrate the gist of text. Ideally, feedback providers should possess 3D and image editing skills to create reference images that can effectively describe what they have in mind. However, they often lack such skills, so they have to resort to sketches or online images that might not match well with the current 3D design. To address this, we introduce\n            <jats:italic>MemoVis<\/jats:italic>\n            , a text editor interface that assists feedback providers in creating reference images with generative AI driven by the feedback comments. First, a novel real-time viewpoint suggestion feature, based on a vision-language foundation model, helps feedback providers anchor a comment with a camera viewpoint. Second, given a camera viewpoint, we introduce three types of image modifiers based on pre-trained 2D generative models to turn a text comment into an updated version of the 3D scene from that viewpoint. We conducted a within-subjects study with\n            <jats:inline-formula content-type=\"math\/tex\">\n              <jats:tex-math notation=\"LaTeX\" version=\"MathJax\">\\(14\\)<\/jats:tex-math>\n            <\/jats:inline-formula>\n            feedback providers, demonstrating the effectiveness of MemoVis. The quality and explicitness of the companion images were evaluated by another eight participants with prior 3D design experience.\n          <\/jats:p>","DOI":"10.1145\/3694681","type":"journal-article","created":{"date-parts":[[2024,9,4]],"date-time":"2024-09-04T12:11:39Z","timestamp":1725451899000},"page":"1-41","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":9,"title":["MemoVis: A GenAI-Powered Tool for Creating Companion Reference Images for 3D Design Feedback"],"prefix":"10.1145","volume":"31","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7179-0861","authenticated-orcid":false,"given":"Chen","family":"Chen","sequence":"first","affiliation":[{"name":"Computer Science and Engineering, University of California San Diego, La Jolla, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9234-9960","authenticated-orcid":false,"given":"Cuong","family":"Nguyen","sequence":"additional","affiliation":[{"name":"Adobe Research, San Francisco, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7984-8252","authenticated-orcid":false,"given":"Thibault","family":"Groueix","sequence":"additional","affiliation":[{"name":"Adobe Research, San Francisco, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3996-6588","authenticated-orcid":false,"given":"Vladimir G.","family":"Kim","sequence":"additional","affiliation":[{"name":"Adobe Research, Seattle, WA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3457-4227","authenticated-orcid":false,"given":"Nadir","family":"Weibel","sequence":"additional","affiliation":[{"name":"Computer Science and Engineering, University of California San Diego, La Jolla, CA, USA"}]}],"member":"320","published-online":{"date-parts":[[2024,11,10]]},"reference":[{"key":"e_1_3_3_2_2","unstructured":"Google. 2015. DeepDream - A Code Example for Visualizing Neural Networks. Retrieved August 8 2023 from https:\/\/ai.googleblog.com\/2015\/07\/deepdream-code-example-for-visualizing.html"},{"key":"e_1_3_3_3_2","unstructured":"OpenAI. 2020. Language Models are Few-shot Learners. Retrieved January 9 2023 from https:\/\/openai.com\/research\/language-models-are-few-shot-learners"},{"key":"e_1_3_3_4_2","unstructured":"OpenAI. 2021. CLIP: Connecting Text and Images. Retrieved December 17 2023 from https:\/\/openai.com\/research\/clip"},{"key":"e_1_3_3_5_2","unstructured":"Midjourney. 2022. Retrieved August 8 2023 from https:\/\/www.midjourney.com"},{"key":"e_1_3_3_6_2","unstructured":"Babylon.js. 2023. Retrieved December 17 2023 from https:\/\/www.babylonjs.com"},{"key":"e_1_3_3_7_2","unstructured":"Lexica. 2023. Retrieved December 17 2023 from https:\/\/lexica.art"},{"key":"e_1_3_3_8_2","unstructured":"Polycount. 2023. Retrieved December 17 2023 from https:\/\/polycount.com"},{"key":"e_1_3_3_9_2","unstructured":"Polycount. 2023. 3D Arts Showcases and Critiques. Retrieved January 27 2024 from https:\/\/polycount.com\/categories\/3d-art-showcase-critiques"},{"key":"e_1_3_3_10_2","unstructured":"2023. Vizcom. 2023. The Next Generation of Product Visualization. Retrieved December 17 2023 from https:\/\/www.vizcom.ai"},{"key":"e_1_3_3_11_2","unstructured":"Adobe. 2023. Adobe Firefly. Retrieved December 17 2023 from https:\/\/www.adobe.com\/products\/firefly.html"},{"key":"e_1_3_3_12_2","unstructured":"Adobe. 2023. Adobe Photoshop. Retrieved December 17 2023 from https:\/\/www.adobe.com\/products\/photoshop.html"},{"key":"e_1_3_3_13_2","unstructured":"Adobe. 2023. Generative Fill Feature from Adobe Photoshop. Retrieved December 17 2023 from https:\/\/www.adobe.com\/products\/photoshop\/generative-fill.html"},{"key":"e_1_3_3_14_2","unstructured":"Adobe. 2023. How to Use Lasso Tool in Adobe Photoshop. Retrieved December 17 2023 from https:\/\/www.adobe.com\/products\/photoshop\/lasso-tool.html"},{"key":"e_1_3_3_15_2","unstructured":"Adobe. 2024. Tap into the Power of AI Photo Editing. Retrieved August 4 2024 from https:\/\/www.adobe.com\/products\/photoshop\/ai.html"},{"key":"e_1_3_3_16_2","unstructured":"Stable Diffusion Art. 2023. How to Remove Undesirable Objects with AI Inpainting. Retrieved December 17 2023 from https:\/\/stable-diffusion-art.com\/how-to-remove-a-person-with-ai-inpainting\/"},{"key":"e_1_3_3_17_2","unstructured":"Autodesk. 2023. Add Annotation - Autodesk Viewer Guide. Retrieved August 4 2024 from https:\/\/help.autodesk.com\/view\/adskviewer\/enu\/?guid=ADSKVIEWER_Help_AddAnnotations_html"},{"key":"e_1_3_3_18_2","unstructured":"Yogesh Balaji Seungjun Nah Xun Huang Arash Vahdat Jiaming Song Qinsheng Zhang Karsten Kreis Miika Aittala Timo Aila Samuli Laine Bryan Catanzaro Tero Karras and Ming-Yu Liu. 2023. eDiff-I: Text-to-image diffusion models with an ensemble of expert denoisers. arXiv:2211.01324. Retrieved from https:\/\/arxiv.org\/abs\/2211.01324"},{"key":"e_1_3_3_19_2","doi-asserted-by":"publisher","DOI":"10.1115\/1.4037109"},{"key":"e_1_3_3_20_2","first-page":"18","volume-title":"Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems (CHI \u201923)","author":"Bennett Dan","year":"2023","unstructured":"Dan Bennett, Oussama Metatla, Anne Roudaut, and Elisa D. Mekler. 2023. How Does HCI Understand Human Agency and Autonomy?. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems (CHI \u201923). ACM, New York, NY, Article 375, 18 pages. DOI: 10.1145\/3544548.3580651"},{"key":"e_1_3_3_21_2","unstructured":"Shariq Farooq Bhat Niloy J. Mitra and Peter Wonka. 2023. LooseControl: Lifting controlnet for generalized depth conditioning. arXiv:2312.03079. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2312.03079"},{"key":"e_1_3_3_22_2","first-page":"133","volume-title":"Analyzing and Interpreting Qualitative Data: After the Interview","author":"Bingham Andrea J.","year":"2022","unstructured":"Andrea J. Bingham and Patricia Witkowsky. 2022. Deductive and inductive approaches to qualitative data analysis. In Analyzing and Interpreting Qualitative Data: After the Interview. C. Vanover, P. Mihas, and J. Salda\u00f1a (Eds.), SAGE Publications, 133\u2013146."},{"key":"e_1_3_3_23_2","doi-asserted-by":"publisher","DOI":"10.1068\/p2897"},{"key":"e_1_3_3_24_2","doi-asserted-by":"publisher","DOI":"10.1037\/13620-004"},{"key":"e_1_3_3_25_2","unstructured":"Raluca Budiu. 2013. Interaction Cost. Retrieved January 16 2023 from https:\/\/www.nngroup.com\/articles\/interaction-cost-definition"},{"key":"e_1_3_3_26_2","doi-asserted-by":"crossref","first-page":"167","DOI":"10.1145\/1111411.1111442","volume-title":"Proceedings of the 2006 Symposium on Interactive 3D Graphics and Games (I3D \u201906)","author":"Burtnyk Nicolas","year":"2006","unstructured":"Nicolas Burtnyk, Azam Khan, George Fitzmaurice, and Gordon Kurtenbach. 2006. ShowMotion: Camera Motion based 3D Design Review. In Proceedings of the 2006 Symposium on Interactive 3D Graphics and Games (I3D \u201906). ACM, New York, NY, 167\u2013174. DOI: 10.1145\/1111411.1111442"},{"key":"e_1_3_3_27_2","first-page":"1","volume-title":"Proceedings of the ACM Collective Intelligence Conference (CI \u201923)","author":"Cai Alice","year":"2023","unstructured":"Alice Cai, Steven R. Rick, Jennifer L. Heyman, Yanxia Zhang, Alexandre Filipowicz, Matthew Hong, Matt Klenk, and Thomas Malone. 2023. DesignAID: Using Generative AI and Semantic Diversity for Design Inspiration. In Proceedings of the ACM Collective Intelligence Conference (CI \u201923). ACM, New York, NY, 1\u201311. DOI: 10.1145\/3582269.3615596"},{"key":"e_1_3_3_28_2","unstructured":"Computer Careers. 2023. Is 3D Modeling Hard? And Other Things You Need To Know. Retrieved January 26 2024 from https:\/\/www.computercareers.org\/is-3d-modeling-hard\/"},{"key":"e_1_3_3_29_2","unstructured":"Xiang \u2019Anthony\u2019 Chen Jeff Burke Ruofei Du Matthew K. Hong Jennifer Jacobs Philippe Laban Dingzeyu Li Nanyun Peng Karl D. D. Willis Chien-Sheng Wu and Bolei Zhou. 2023. Next steps for human-centered generative AI: A technical perspective. arXiv:2306.15774. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2306.15774"},{"key":"e_1_3_3_30_2","first-page":"25","volume-title":"Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201924)","author":"Choi DaEun","year":"2024","unstructured":"DaEun Choi, Sumin Hong, Jeongeon Park, John Joon Young Chung, and Juho Kim. 2024. CreativeConnect: Supporting Reference Recombination for Graphic Design Ideation with Generative AI. In Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201924). ACM, New York, NY, Article 1055, 25 pages. DOI: 10.1145\/3613904.3642794"},{"key":"e_1_3_3_31_2","first-page":"539","volume-title":"Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition","volume":"1","author":"Chopra Sumit","year":"2005","unstructured":"Sumit Chopra, Raia Hadsell, and Yann LeCun. 2005. Learning a Similarity Metric Discriminatively, with Application to Face Verification. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Vol. 1, 539\u2013546. DOI: 10.1109\/cvpr.2005.202"},{"key":"e_1_3_3_32_2","volume-title":"Statistical Power Analysis for Behavioral Sciences","author":"Cohen H.","year":"1988","unstructured":"H. Cohen. 1988. Statistical Power Analysis for Behavioral Sciences. Lawrence Erlbaum Associates, Hillsdale, NJ."},{"key":"e_1_3_3_33_2","first-page":"2","volume-title":"Proceedings of the ACM SIGGRAPH 2022 Posters (SIGGRAPH \u201922)","author":"Dukor Obumneme Stanley","year":"2022","unstructured":"Obumneme Stanley Dukor, S. Mahdi H. Miangoleh, Mahesh Kumar Krishna Reddy, Long Mai, and Ya\u011fi\u0307z Aksoy. 2022. Interactive Editing of Monocular Depth. In Proceedings of the ACM SIGGRAPH 2022 Posters (SIGGRAPH \u201922). ACM, New York, NY, Article 52, 2 pages. DOI: 10.1145\/3532719.3543235"},{"key":"e_1_3_3_34_2","first-page":"93","volume-title":"Proceedings of the Conference on Artificial Intelligence in Education: Building Technology Rich Learning Contexts That Work","author":"Easterday Matthew W.","year":"2007","unstructured":"Matthew W. Easterday, Vincent Aleven, and Richard Scheines. 2007. \u2019Tis Better to Construct than to Receive? The Effects of Diagram Tools on Causal Reasoning. In Proceedings of the Conference on Artificial Intelligence in Education: Building Technology Rich Learning Contexts That Work. IOS Press, 93\u2013100."},{"key":"e_1_3_3_35_2","doi-asserted-by":"publisher","DOI":"10.1177\/1473871611413180"},{"key":"e_1_3_3_36_2","first-page":"10","volume-title":"Proceedings of the 35th Annual ACM Symposium on User Interface Software and Technology (UIST \u201922)","author":"Evirgen Noyan","year":"2022","unstructured":"Noyan Evirgen and Xiang \u2019Anthony\u2019 Chen. 2022. GANzilla: User-Driven Direction Discovery in Generative Adversarial Networks. In Proceedings of the 35th Annual ACM Symposium on User Interface Software and Technology (UIST \u201922). ACM, New York, NY, Article 75, 10 pages. DOI: 10.1145\/3526113.3545638"},{"key":"e_1_3_3_37_2","first-page":"15","volume-title":"Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems (CHI \u201923)","author":"Evirgen Noyan","year":"2023","unstructured":"Noyan Evirgen and Xiang \u2019Anthony Chen. 2023. GANravel: User-Driven Direction Disentanglement in Generative Adversarial Networks. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems (CHI \u201923). ACM, New York, NY, Article 19, 15 pages. DOI: 10.1145\/3544548.3581226"},{"issue":"1","key":"e_1_3_3_38_2","doi-asserted-by":"crossref","first-page":"243","DOI":"10.1037\/0033-295X.112.1.243","article-title":"Information along Contours and Object Boundaries","volume":"112","author":"Feldman Jacob","year":"2005","unstructured":"Jacob Feldman and Manish Singh. 2005. Information along Contours and Object Boundaries. Psychological Review 112, 1 (2005), 243.","journal-title":"Psychological Review"},{"key":"e_1_3_3_39_2","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2023.3327168"},{"key":"e_1_3_3_40_2","unstructured":"Sarah Gibbons. 2016. Design Critiques: Encourage a Positive Culture to Improve Products. Retrieved August 2 2023 from https:\/\/www.nngroup.com\/articles\/design-critiques"},{"key":"e_1_3_3_41_2","doi-asserted-by":"crossref","DOI":"10.4135\/9781412983419","volume-title":"ANOVA: Repeated Measures","author":"Girden Ellen R.","year":"1992","unstructured":"Ellen R. Girden. 1992. ANOVA: Repeated Measures. Vol. 84. Sage University Paper Series."},{"key":"e_1_3_3_42_2","unstructured":"Ian J. Goodfellow Jean Pouget-Abadie Mehdi Mirza Bing Xu David Warde-Farley Sherjil Ozair Aaron Courville and Yoshua Bengio. 2014. Generative adversarial networks. arXiv:1406.2661. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.1406.2661"},{"key":"e_1_3_3_43_2","unstructured":"Yaru Hao Zewen Chi Li Dong and Furu Wei. 2023. Optimizing prompts for text-to-image generation. arXiv:2212.09611. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2212.09611"},{"key":"e_1_3_3_44_2","doi-asserted-by":"publisher","DOI":"10.3389\/frai.2022.976235"},{"key":"e_1_3_3_45_2","doi-asserted-by":"crossref","first-page":"87","DOI":"10.1145\/1518701.1518717","volume-title":"Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI \u201909)","author":"Herring Scarlett R.","year":"2009","unstructured":"Scarlett R. Herring, Chia-Chen Chang, Jesse Krantzler, and Brian P. Bailey. 2009. Getting Inspired! Understanding How and Why Examples are Used in Creative Design Practice. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI \u201909). ACM, New York, NY, 87\u201396. DOI: 10.1145\/1518701.1518717"},{"key":"e_1_3_3_46_2","first-page":"2434","volume-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","author":"Holinaty Josh","year":"2021","unstructured":"Josh Holinaty, Alec Jacobson, and Fanny Chevalier. 2021. Supporting Reference Imagery for Digital Drawing. In Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2434\u20132442. DOI: 10.1109\/iccvw54120.2021.00276"},{"key":"e_1_3_3_47_2","unstructured":"Xinyu Huang Youcai Zhang Jinyu Ma Weiwei Tian Rui Feng Yuejie Zhang Yaqian Li Yandong Guo and Lei Zhang. 2023. Tag2Text: Guiding vision-language model via image tagging. arXiv:2303.05657. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2303.05657"},{"key":"e_1_3_3_48_2","unstructured":"Apple Inc. 2024. Use Memoji on your iPhone or iPad Pro. Retrieved January 16 2023 from https:\/\/support.apple.com\/en-us\/111115"},{"key":"e_1_3_3_49_2","first-page":"1","volume-title":"Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201918)","author":"Kang Hyeonsu B.","year":"2018","unstructured":"Hyeonsu B. Kang, Gabriel Amoako, Neil Sengupta, and Steven P. Dow. 2018. Paragon: An Online Gallery for Enhancing Design Feedback with Visual Examples. In Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201918). ACM, New York, NY, 1\u201313. DOI: 10.1145\/3173574.3174180"},{"key":"e_1_3_3_50_2","first-page":"4401","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","author":"Karras Tero","year":"2019","unstructured":"Tero Karras, Samuli Laine, and Timo Aila. 2019. A Style-Based Generator Architecture for Generative Adversarial Networks. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 4401\u20134410."},{"key":"e_1_3_3_51_2","first-page":"8110","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","author":"Karras Tero","year":"2020","unstructured":"Tero Karras, Samuli Laine, Miika Aittala, Janne Hellsten, Jaakko Lehtinen, and Timo Aila. 2020. Analyzing and Improving the Image Quality of Stylegan. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 8110\u20138119."},{"key":"e_1_3_3_52_2","first-page":"625","volume-title":"Proceedings of the Conference on Designing Interactive Systems (DIS \u201917)","author":"Kim Chang Min","year":"2017","unstructured":"Chang Min Kim, Hyeon-Beom Yi, Ji-Won Nam, and Geehyuk Lee. 2017. Applying Real-Time Text on Instant Messaging for a Rapid and Enriched Conversation Experience. In Proceedings of the Conference on Designing Interactive Systems (DIS \u201917). ACM, New York, NY, 625\u2013629. DOI: 10.1145\/3064663.3064679"},{"key":"e_1_3_3_53_2","doi-asserted-by":"crossref","unstructured":"Alexander Kirillov Eric Mintun Nikhila Ravi Hanzi Mao Chloe Rolland Laura Gustafson Tete Xiao Spencer Whitehead Alexander C. Berg Wan-Yen Lo Piotr Doll\u00e1r and Ross Girshick. 2023. Segment anything. arXiv:2304.02643. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2304.02643","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"e_1_3_3_54_2","first-page":"919","volume-title":"Proceedings of the 28th International Conference on Intelligent User Interfaces (IUI \u201923)","author":"Ko Hyung-Kwon","year":"2023","unstructured":"Hyung-Kwon Ko, Gwanmo Park, Hyeon Jeon, Jaemin Jo, Juho Kim, and Jinwook Seo. 2023. Large-Scale Text-to-Image Generation Models for Visual Artists\u2019 Creative Works. In Proceedings of the 28th International Conference on Intelligent User Interfaces (IUI \u201923). ACM, New York, NY, 919\u2013933. DOI: 10.1145\/3581641.3584078"},{"key":"e_1_3_3_55_2","unstructured":"Tanner Kohler. 2022. Three Methods to Increase User Autonomy in UX Design. Retrieved January 16 2023 from https:\/\/www.nngroup.com\/articles\/increase-user-autonomy"},{"issue":"4","key":"e_1_3_3_56_2","first-page":"12","article-title":"Sequential Gallery for Interactive Visual Design Optimization","volume":"39","author":"Koyama Yuki","year":"2020","unstructured":"Yuki Koyama, Issei Sato, and Masataka Goto. 2020. Sequential Gallery for Interactive Visual Design Optimization. ACM Transactions on Graphics 39, 4 (Aug 2020), Article 88, 12 pages. DOI: https:\/\/doi.org\/10.1145\/3386569.3392444","journal-title":"ACM Transactions on Graphics"},{"key":"e_1_3_3_57_2","doi-asserted-by":"publisher","DOI":"10.1145\/3072959.3073598"},{"key":"e_1_3_3_58_2","first-page":"4627","volume-title":"Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201917)","author":"Krause Markus","year":"2017","unstructured":"Markus Krause, Tom Garncarz, JiaoJiao Song, Elizabeth M. Gerber, Brian P. Bailey, and Steven P. Dow. 2017. Critique Style Guide: Improving Crowdsourced Design Feedback with a Natural Language Model. In Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201917). ACM, New York, NY, 4627\u20134639. DOI: 10.1145\/3025453.3025883"},{"key":"e_1_3_3_59_2","doi-asserted-by":"publisher","DOI":"10.1109\/tvcg.2008.109"},{"key":"e_1_3_3_60_2","first-page":"264","volume-title":"Proceedings of the 28th International Conference on Intelligent User Interfaces (IUI \u201923)","author":"Lawton Tomas","year":"2023","unstructured":"Tomas Lawton, Francisco J. Ibarrola, Dan Ventura, and Kazjon Grace. 2023. Drawing with Reframer: Emergence and Control in Co-Creative AI. In Proceedings of the 28th International Conference on Intelligent User Interfaces (IUI \u201923). ACM, New York, NY, 264\u2013277. DOI: 10.1145\/3581641.3584095"},{"key":"e_1_3_3_61_2","first-page":"18","volume-title":"Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201924)","author":"Lee Seung Won","year":"2024","unstructured":"Seung Won Lee, Tae Hee Jo, Semin Jin, Jiin Choi, Kyungwon Yun, Sergio Bromberg, Seonghoon Ban, and Kyung Hoon Hyun. 2024. The Impact of Sketch-guided vs. Prompt-guided 3D Generative AIs on the Design Exploration Process. In Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201924). ACM, New York, NY, Article 1057, 18 pages. DOI: 10.1145\/3613904.3642218"},{"key":"e_1_3_3_62_2","unstructured":"Junnan Li Dongxu Li Silvio Savarese and Steven Hoi. 2023. BLIP-2: Bootstrapping language-image pre-training with frozen image encoders and large language models. arXiv:2301.12597. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2301.12597"},{"key":"e_1_3_3_63_2","unstructured":"Junnan Li Dongxu Li Caiming Xiong and Steven Hoi. 2022. BLIP: Bootstrapping language-image pre-training for unified vision-language understanding and generation. arXiv: 2201.12086. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2201.12086"},{"key":"e_1_3_3_64_2","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2111.03186"},{"key":"e_1_3_3_65_2","doi-asserted-by":"publisher","DOI":"10.1115\/1.4003498"},{"key":"e_1_3_3_66_2","first-page":"23","volume-title":"Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201922)","author":"Liu Vivian","year":"2022","unstructured":"Vivian Liu and Lydia B. Chilton. 2022. Design Guidelines for Prompt Engineering Text-to-Image Generative Models. In Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201922). ACM, New York, NY, Article 384, 23 pages. DOI: 10.1145\/3491102.3501825"},{"key":"e_1_3_3_67_2","first-page":"1955","volume-title":"Proceedings of the 2023 ACM Designing Interactive Systems Conference (DIS \u201923)","author":"Liu Vivian","year":"2023","unstructured":"Vivian Liu, Jo Vermeulen, George Fitzmaurice, and Justin Matejka. 2023. 3DALL-E: Integrating Text-to-Image AI in 3D Design Workflows. In Proceedings of the 2023 ACM Designing Interactive Systems Conference (DIS \u201923). ACM, New York, NY, 1955\u20131977. DOI: 10.1145\/3563657.3596098"},{"key":"e_1_3_3_68_2","doi-asserted-by":"publisher","DOI":"10.1145\/2611519"},{"key":"e_1_3_3_69_2","unstructured":"Oscar Ma\u00f1as Pietro Astolfi Melissa Hall Candace Ross Jack Urbanek Adina Williams Aishwarya Agrawal Adriana Romero-Soriano and Michal Drozdzal. 2024. Improving text-to-image consistency via automatic prompt optimization. arXiv:2403.17804. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2403.17804"},{"key":"e_1_3_3_70_2","unstructured":"Meredith Ringel Morris Jascha Sohl-dickstein Noah Fiedel Tris Warkentin Allan Dafoe Aleksandra Faust Clement Farabet and Shane Legg. 2023. Levels of AGI: Operationalizing progress on the path to AGI. arXiv:2311.02462. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2311.02462"},{"key":"e_1_3_3_71_2","first-page":"27","article-title":"Generative and AI Authored Artworks and Copyright Law","volume":"45","author":"Murray Michael D.","year":"2023","unstructured":"Michael D. Murray. 2023. Generative and AI Authored Artworks and Copyright Law. Hastings Communications and Entertainment Law Journal 45 (2023), 27.","journal-title":"Hastings Communications and Entertainment Law Journal"},{"key":"e_1_3_3_72_2","doi-asserted-by":"crossref","first-page":"267","DOI":"10.1145\/3126594.3126659","volume-title":"Proceedings of the 30th Annual ACM Symposium on User Interface Software and Technology (UIST \u201917)","author":"Nguyen Cuong","year":"2017","unstructured":"Cuong Nguyen, Stephen DiVerdi, Aaron Hertzmann, and Feng Liu. 2017. CollaVR: Collaborative In-headset Review for VR Video. In Proceedings of the 30th Annual ACM Symposium on User Interface Software and Technology (UIST \u201917). ACM, New York, NY, 267\u2013277. DOI: 10.1145\/3126594.3126659"},{"key":"e_1_3_3_73_2","doi-asserted-by":"publisher","DOI":"10.1207\/s15327957pspr1003_1"},{"key":"e_1_3_3_74_2","volume-title":"The Design of Everyday Things","author":"Norman Don","year":"2013","unstructured":"Don Norman. 2013. The Design of Everyday Things. Basic Books."},{"key":"e_1_3_3_75_2","first-page":"21","volume-title":"Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201924)","author":"Oh Jeongseok","year":"2024","unstructured":"Jeongseok Oh, Seungju Kim, and Seungjun Kim. 2024. LumiMood: A Creativity Support Tool for Designing the Mood of a 3D Scene. In Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201924). ACM, New York, NY, Article 174, 21 pages. DOI: 10.1145\/3613904.3642440"},{"key":"e_1_3_3_76_2","unstructured":"OpenAI. 2023. Blender Copilot (Blender GPT). Retrieved December 18 2023 from https:\/\/blendermarket.com\/products\/blender-copilot-blendergpt"},{"key":"e_1_3_3_77_2","unstructured":"OpenAI. 2023. DALL.E 3. Retrieved December 17 2023 from https:\/\/openai.com\/dall-e-3"},{"key":"e_1_3_3_78_2","first-page":"933","volume-title":"Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval (SIGIR \u201918)","author":"Otterbacher Jahna","year":"2018","unstructured":"Jahna Otterbacher, Alessandro Checco, Gianluca Demartini, and Paul Clough. 2018. Investigating User Perception of Gender Bias in Image Search: The Role of Sexism. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval (SIGIR \u201918). ACM, New York, NY, 933\u2013936. DOI: 10.1145\/3209978.3210094"},{"key":"e_1_3_3_79_2","doi-asserted-by":"crossref","first-page":"517","DOI":"10.1145\/2984511.2984552","volume-title":"Proceedings of the 29th Annual Symposium on User Interface Software and Technology (UIST \u201916)","author":"Pavel Amy","year":"2016","unstructured":"Amy Pavel, Dan B. Goldman, Bj\u00f6rn Hartmann, and Maneesh Agrawala. 2016. VidCrit: Video-Based Asynchronous Video Review. In Proceedings of the 29th Annual Symposium on User Interface Software and Technology (UIST \u201916). ACM, New York, NY, 517\u2013528. DOI: 10.1145\/2984511.2984552"},{"key":"e_1_3_3_80_2","first-page":"213","volume-title":"Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI \u201996)","author":"Pirolli Peter","year":"1996","unstructured":"Peter Pirolli, Patricia Schank, Marti Hearst, and Christine Diehl. 1996. Scatter\/Gather Browsing Communicates the Topic Structure of a Very Large Text Collection. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI \u201996). ACM, New York, NY, 213\u2013220. DOI: 10.1145\/238386.238489"},{"key":"e_1_3_3_81_2","volume-title":"Proceedings of the International Conference on GraphiCon \u201996","author":"Plemenos Dimitri","year":"1996","unstructured":"Dimitri Plemenos and Madjid Benayada. 1996. Intelligent Display in Scene Modelling. New Techniques to Automatically Compute Good Views. In Proceedings of the International Conference on GraphiCon \u201996."},{"key":"e_1_3_3_82_2","unstructured":"Alec Radford Jong Wook Kim Chris Hallacy Aditya Ramesh Gabriel Goh Sandhini Agarwal Girish Sastry Amanda Askell Pamela Mishkin Jack Clark Gretchen Krueger and Ilya Sutskever. 2021. Learning transferable visual models from natural language supervision. arXiv:2103.00020. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2103.00020"},{"key":"e_1_3_3_83_2","doi-asserted-by":"crossref","unstructured":"Shwetha Rajaram Nels Numan Balasaravanan Thoravi Kumaravel Nicolai Marquardt and Andrew D. Wilson. 2024. BlendScape: Enabling unified and personalized video-conferencing environments through generative AI. arXiv: 2403.13947. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2403.13947","DOI":"10.1145\/3654777.3676326"},{"key":"e_1_3_3_84_2","first-page":"1355","volume-title":"Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems (CHI \u201915)","author":"Robb David A.","year":"2015","unstructured":"David A. Robb, Stefano Padilla, Britta Kalkreuter, and Mike J. Chantler. 2015. Crowdsourced Feedback with Imagery Rather Than Text: Would Designers Use It?. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems (CHI \u201915). ACM, New York, NY, 1355\u20131364. DOI: 10.1145\/2702123.2702470"},{"key":"e_1_3_3_85_2","doi-asserted-by":"crossref","unstructured":"Robin Rombach Andreas Blattmann Dominik Lorenzm Patrick Esser and Bj\u00f6rn Ommer. 2022. High-resolution image synthesis with latent diffusion models. arXiv:2112.10752. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2112.10752","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"e_1_3_3_86_2","doi-asserted-by":"publisher","DOI":"10.1126\/science.adi0656"},{"key":"e_1_3_3_87_2","doi-asserted-by":"publisher","DOI":"10.1145\/2019627.2019628"},{"key":"e_1_3_3_88_2","unstructured":"Mohamad Shahbazi Liesbeth Claessens Michael Niemeyer Edo Collins Alessio Tonioni Luc Van Gool and Federico Tombari. 2024. InseRF: Text-driven generative object insertion in neural 3D scenes. arXiv:2401.05335. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2401.05335"},{"key":"e_1_3_3_89_2","doi-asserted-by":"publisher","DOI":"10.1093\/biomet\/52.3-4.591"},{"key":"e_1_3_3_90_2","unstructured":"Kihoon Son DaEun Choi Tae Soo Kim Young-Ho Kim and Juho Kim. 2023. GenQuery: Supporting expressive visual search with generative models. arXiv: 2310.01287. Retrieved from https:\/\/arxiv.org\/abs\/2310.01287"},{"key":"e_1_3_3_91_2","doi-asserted-by":"publisher","DOI":"10.1145\/1592440.1592443"},{"key":"e_1_3_3_92_2","doi-asserted-by":"crossref","unstructured":"Christian Szegedy Wei Liu Yangqing Jia Pierre Sermanet Scott Reed Dragomir Anguelov Dumitru Erhan Vincent Vanhoucke and Andrew Rabinovich. 2014. Going deeper with convolutions. arXiv:1409.4842. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.1409.4842","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"e_1_3_3_93_2","unstructured":"Promatics Technologies. 2017. An Overview of Asynchronous Design Feedback and Its Benefits. Retrieved December 16 2023 from https:\/\/medium.com\/@promatics\/22a6b97b33f0"},{"key":"e_1_3_3_94_2","unstructured":"Autodesk TinkerCAD. 2020. Annotate Tinkercad Designs with 3D Notes. Retrieved August 7 2023 from https:\/\/www.tinkercad.com\/blog\/annotate-tinkercad-designs-with-3d-notes"},{"key":"e_1_3_3_95_2","doi-asserted-by":"publisher","DOI":"10.1111\/j.1467-8659.2009.01412.x"},{"key":"e_1_3_3_96_2","doi-asserted-by":"crossref","unstructured":"Henrik Voigt Jan Hombeck Monique Meuschke Kai Lawonn and Sina Zarrie\u00df. 2023. Paparazzi: A deep dive into the capabilities of language and vision models for grounding viewpoint descriptions. arXiv:2302.10282. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2302.10282","DOI":"10.18653\/v1\/2023.findings-eacl.62"},{"key":"e_1_3_3_97_2","doi-asserted-by":"crossref","first-page":"136","DOI":"10.1145\/3563657.3596072","volume-title":"Proceedings of the 2023 ACM Designing Interactive Systems Conference (DIS \u201923)","author":"Wan Qian","year":"2023","unstructured":"Qian Wan and Zhicong Lu. 2023. GANCollage: A GAN-Driven Digital Mood Board to Facilitate Ideation in Creativity Support. In Proceedings of the 2023 ACM Designing Interactive Systems Conference (DIS \u201923). ACM, New York, NY, 136\u2013146. DOI: 10.1145\/3563657.3596072"},{"key":"e_1_3_3_98_2","first-page":"1805","volume-title":"Proceedings of the Design Society","volume":"3","author":"Wang Da","year":"2023","unstructured":"Da Wang and Ji Han. 2023. Exploring the Impact of Generative Stimuli on the Creativity of Designers in Combinational Design. Proceedings of the Design Society 3 (2023), 1805\u20131814. DOI: 10.1017\/pds.2023.181"},{"key":"e_1_3_3_99_2","unstructured":"Qian Wang Biao Zhang Michael Birsak and Peter Wonka. 2023. InstructEdit: Improving automatic masks for diffusion-based image editing with user instructions. arXiv:2305.18047. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2305.18047"},{"key":"e_1_3_3_100_2","first-page":"21","volume-title":"Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201924)","author":"Wang Zhijie","year":"2024","unstructured":"Zhijie Wang, Yuheng Huang, Da Song, Lei Ma, and Tianyi Zhang. 2024. PromptCharm: Text-to-Image Generation through Multi-modal Prompting and Refinement. In Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201924). ACM, New York, NY, Article 185, 21 pages. DOI: 10.1145\/3613904.3642803"},{"key":"e_1_3_3_101_2","first-page":"695","volume-title":"Proceedings of the 28th International Conference on Intelligent User Interfaces (IUI \u201923)","author":"Warner Jeremy","year":"2023","unstructured":"Jeremy Warner, Amy Pavel, Tonya Nguyen, Maneesh Agrawala, and Bj\u00f6ern Hartmann. 2023. SlideSpecs: Automatic and interactive presentation feedback collation. In Proceedings of the 28th International Conference on Intelligent User Interfaces (IUI \u201923). ACM, New York, NY, 695\u2013709. DOI: 10.1145\/3581641.3584035"},{"key":"e_1_3_3_102_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.autcon.2019.03.018"},{"key":"e_1_3_3_103_2","unstructured":"Chenfei Wu Shengming Yin Weizhen Qi Xiaodong Wang Zecheng Tang and Nan Duan. 2023. Visual ChatGPT: Talking drawing and editing with visual foundation models. arXiv:2303.04671. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2303.04671"},{"key":"e_1_3_3_104_2","volume-title":"Proceedings of IEEE International Conference on Computer Vision","author":"Xie Saining","year":"2015","unstructured":"Saining Xie and Zhuowen Tu. 2015. Holistically-Nested Edge Detection. In Proceedings of IEEE International Conference on Computer Vision. DOI: 10.1109\/iccv.2015.164"},{"key":"e_1_3_3_105_2","unstructured":"Yuewei Yang Xiaoliang Dai Jialiang Wang Peizhao Zhang and Hongbo Zhang. 2023. Efficient quantization strategies for latent diffusion Models. arXiv:2312.05431. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2312.05431"},{"key":"e_1_3_3_106_2","first-page":"15","volume-title":"Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201921)","author":"Zhang Enhao","year":"2021","unstructured":"Enhao Zhang and Nikola Banovic. 2021. Method for Exploring Generative Adversarial Networks (GANs) via Automatically Generated Image Galleries. In Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI \u201921). ACM, New York, NY, Article 76, 15 pages. DOI: 10.1145\/3411764.3445714"},{"key":"e_1_3_3_107_2","first-page":"3836","volume-title":"IEEE International Conference on Computer Vision (ICCV)","author":"Zhang Lvmin","year":"2023","unstructured":"Lvmin Zhang and Maneesh Agrawala. 2023. Adding Conditional Control to Text-to-Image Diffusion Models. In IEEE International Conference on Computer Vision (ICCV), 3836\u20133847. DOI: 10.48550\/arXiv.2302.05543"},{"key":"e_1_3_3_108_2","doi-asserted-by":"crossref","unstructured":"Youcai Zhang Xinyu Huang Jinyu Ma Zhaoyang Li Zhaochuan Luo Yanchun Xie Yuzhuo Qin Tong Luo Yaqian Li Shilong Liu Yandong Guo and Lei Zhang. 2023. Recognize anything: A strong image tagging model. arXiv:2306.03514. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.2306.03514","DOI":"10.1109\/CVPRW63382.2024.00179"},{"key":"e_1_3_3_109_2","first-page":"11127","volume-title":"Proceedings of the 37th International Conference onNeural Information Processing Systems","author":"Zhao Shihao","year":"2023","unstructured":"Shihao Zhao, Dongdong Chen, Yen-Chun Chen, Jianmin Bao, Shaozhe Hao, Lu Yuan, and Kwan-Yee K. Wong. 2023. Uni-ControlNet: All-in-One Control to Text-to-Image Diffusion Models. In Proceedings of the 37th International Conference onNeural Information Processing Systems, 11127\u201311150. DOI: 10.48550\/arXiv.2305.16322"},{"key":"e_1_3_3_110_2","unstructured":"Daniel M. Ziegler Nisan Stiennon Jeffrey Wu Tom B. Brown Alec Radford Dario Amodei Paul Christiano and Geoffrey Irving. 2020. Fine-tuning language models from human preferences. arXiv:1909.08593v2. Retrieved from https:\/\/doi.org\/10.48550\/arXiv.1909.08593"}],"container-title":["ACM Transactions on Computer-Human Interaction"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3694681","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3694681","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:05:47Z","timestamp":1750291547000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3694681"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,31]]},"references-count":109,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2024,10,31]]}},"alternative-id":["10.1145\/3694681"],"URL":"https:\/\/doi.org\/10.1145\/3694681","relation":{},"ISSN":["1073-0516","1557-7325"],"issn-type":[{"value":"1073-0516","type":"print"},{"value":"1557-7325","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10,31]]},"assertion":[{"value":"2024-01-29","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-08-19","order":2,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-11-10","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}