{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T18:52:39Z","timestamp":1776106359866,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":25,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,10,29]],"date-time":"2023-10-29T00:00:00Z","timestamp":1698537600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,10,29]]},"DOI":"10.1145\/3586182.3616623","type":"proceedings-article","created":{"date-parts":[[2023,10,27]],"date-time":"2023-10-27T19:50:32Z","timestamp":1698436232000},"page":"1-3","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":10,"title":["Generative Facial Expressions and Eye Gaze Behavior from Prompts for Multi-Human-Robot Interaction"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0009-2424-6527","authenticated-orcid":false,"given":"Gabriel J","family":"Serfaty","sequence":"first","affiliation":[{"name":"University of Michigan, United States"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-8673-2334","authenticated-orcid":false,"given":"Virgil O","family":"Barnard","sequence":"additional","affiliation":[{"name":"Open Innovation Center, Riverside Research, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9448-9024","authenticated-orcid":false,"given":"Joseph P","family":"Salisbury","sequence":"additional","affiliation":[{"name":"Open Innovation Center, Riverside Research, United States"}]}],"member":"320","published-online":{"date-parts":[[2023,10,29]]},"reference":[{"key":"e_1_3_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.leaqua.2019.101337"},{"key":"e_1_3_2_2_2_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1803916115"},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.1037\/0033-2909.100.1.78"},{"key":"e_1_3_2_2_4_1","doi-asserted-by":"publisher","DOI":"10.5465\/amj.2017.1507"},{"key":"e_1_3_2_2_5_1","doi-asserted-by":"publisher","DOI":"10.5898\/JHRI.6.1.Admoni"},{"key":"e_1_3_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICHR.2006.321322"},{"key":"e_1_3_2_2_7_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.artint.2005.03.005"},{"key":"e_1_3_2_2_8_1","unstructured":"B. Mutlu T. Shiwa T. Kanda H. Ishiguro and N. Hagita \u201cFooting In Human-Robot Conversations: How Robots Might Shape Participant Roles Using Gaze Cues\u201d."},{"key":"e_1_3_2_2_9_1","volume-title":"Improving language understanding by generative pre-training","author":"Radford A.","year":"2018","unstructured":"A. Radford, K. Narasimhan, T. Salimans, and I. Sutskever, \u201cImproving language understanding by generative pre-training,\u201d 2018."},{"issue":"8","key":"e_1_3_2_2_10_1","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford A.","year":"2019","unstructured":"A. Radford, J. Wu, R. Child, D. Luan, D. Amodei, and I. Sutskever, \u201cLanguage models are unsupervised multitask learners,\u201d OpenAI blog, vol. 1, no. 8, p. 9, 2019.","journal-title":"OpenAI blog"},{"key":"e_1_3_2_2_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3568294.3580040"},{"key":"e_1_3_2_2_12_1","volume-title":"Feb. 04, 2023. https:\/\/arxiv.org\/abs\/2302.02083","author":"Kosinski M.","year":"2023","unstructured":"M. Kosinski, \u201cTheory of Mind May Have Spontaneously Emerged in Large Language Models,\u201d arXiv.org, Feb. 04, 2023. https:\/\/arxiv.org\/abs\/2302.02083v3 (accessed Jun. 19, 2023)."},{"key":"e_1_3_2_2_13_1","unstructured":"B. Zhang and H. Soh \u201cLarge Language Models as Zero-Shot Human Models for Human-Robot Interaction\u201d."},{"key":"e_1_3_2_2_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544549.3585602"},{"key":"e_1_3_2_2_15_1","volume-title":"Oct. 07, 2021. Accessed","author":"Xie B.","year":"2023","unstructured":"B. Xie and C. H. Park, \u201cA MultiModal Social Robot Toward Personalized Emotion Interaction.\u201d arXiv, Oct. 07, 2021. Accessed: May 25, 2023. [Online]. Available: http:\/\/arxiv.org\/abs\/2110.05186"},{"key":"e_1_3_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3560815"},{"key":"e_1_3_2_2_17_1","volume-title":"Feb. 21, 2023. Accessed","author":"White J.","year":"2023","unstructured":"J. White , \u201cA Prompt Pattern Catalog to Enhance Prompt Engineering with ChatGPT.\u201d arXiv, Feb. 21, 2023. Accessed: Jun. 19, 2023. [Online]. Available: http:\/\/arxiv.org\/abs\/2302.11382"},{"key":"e_1_3_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-34584-5_9"},{"key":"e_1_3_2_2_19_1","unstructured":"\u201cfurhat-remote-api: Furhat Remote API.\u201d"},{"key":"e_1_3_2_2_20_1","volume-title":"Remote API - Furhat Developer Docs","year":"2023","unstructured":"\u201cRemote API - Furhat Developer Docs.\u201d https:\/\/docs.furhat.io\/remote-api\/ (accessed Jun. 19, 2023)."},{"key":"e_1_3_2_2_21_1","volume-title":"Accessed","year":"2023","unstructured":"OpenAI, \u201copenai: Python client library for the OpenAI API.\u201d Accessed: Jun. 19, 2023. [OS Independent]. Available: https:\/\/github.com\/openai\/openai-python"},{"key":"e_1_3_2_2_22_1","volume-title":"OpenAI API","year":"2023","unstructured":"\u201cOpenAI API.\u201d https:\/\/platform.openai.com\/docs\/ (accessed Jun. 19, 2023)."},{"key":"e_1_3_2_2_23_1","volume-title":"Azure OpenAI Service \u2013 Advanced Language Models | Microsoft Azure","year":"2023","unstructured":"\u201cAzure OpenAI Service \u2013 Advanced Language Models | Microsoft Azure.\u201d https:\/\/azure.microsoft.com\/en-us\/products\/cognitive-services\/openai-service (accessed Apr. 04, 2023)."},{"key":"e_1_3_2_2_24_1","volume-title":"Google Cloud. https:\/\/cloud.google.com\/speech-to-text (accessed","year":"2023","unstructured":"\u201cSpeech-to-Text: Automatic Speech Recognition,\u201d Google Cloud. https:\/\/cloud.google.com\/speech-to-text (accessed Jun. 19, 2023)."},{"key":"e_1_3_2_2_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3568294.3580049"}],"event":{"name":"UIST '23: The 36th Annual ACM Symposium on User Interface Software and Technology","location":"San Francisco CA USA","acronym":"UIST '23","sponsor":["SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques","SIGCHI ACM Special Interest Group on Computer-Human Interaction"]},"container-title":["Adjunct Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3586182.3616623","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3586182.3616623","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:36:41Z","timestamp":1750178201000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3586182.3616623"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,29]]},"references-count":25,"alternative-id":["10.1145\/3586182.3616623","10.1145\/3586182"],"URL":"https:\/\/doi.org\/10.1145\/3586182.3616623","relation":{},"subject":[],"published":{"date-parts":[[2023,10,29]]},"assertion":[{"value":"2023-10-29","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}