{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T20:48:28Z","timestamp":1776113308359,"version":"3.50.1"},"reference-count":84,"publisher":"Association for Computing Machinery (ACM)","issue":"1","license":[{"start":{"date-parts":[[2025,3,3]],"date-time":"2025-03-03T00:00:00Z","timestamp":1740960000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Science Foundation of China","doi-asserted-by":"crossref","award":["62202407"],"award-info":[{"award-number":["62202407"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Hong Kong RGC grants","award":["14214022, 14207123, 14201924, C4072-21G, and T43-513\/23-N"],"award-info":[{"award-number":["14214022, 14207123, 14201924, C4072-21G, and T43-513\/23-N"]}]},{"DOI":"10.13039\/501100006374","name":"National Science Foundation","doi-asserted-by":"publisher","award":["CNS-1943396"],"award-info":[{"award-number":["CNS-1943396"]}],"id":[{"id":"10.13039\/501100006374","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["Proc. ACM Interact. Mob. Wearable Ubiquitous Technol."],"published-print":{"date-parts":[[2025,3,3]]},"abstract":"<jats:p>Social interactions are fundamental to human life. The recent emergence of large language models (LLMs)-based virtual assistants has demonstrated their potential to revolutionize human interactions and lifestyles. However, existing assistive systems mainly provide reactive services to individual users, rather than offering in-situ assistance during live social interactions with conversational partners. In this study, we introduce SocialMind, the first LLM-based proactive AR social assistive system that provides users with in-situ social assistance. SocialMind employs human-like perception leveraging multi-modal sensors to extract both verbal and nonverbal cues, social factors, and implicit personas, incorporating these social cues into LLM reasoning for social suggestion generation. Additionally, SocialMind employs a multi-tier collaborative generation strategy and proactive update mechanism to display social suggestions on Augmented Reality (AR) glasses, ensuring that suggestions are timely provided to users without disrupting the natural flow of conversation. Evaluations on three public datasets and a user study with 20 participants show that SocialMind achieves 38.3% higher engagement compared to baselines, and 95% of participants are willing to use SocialMind in their live social interactions.<\/jats:p>","DOI":"10.1145\/3712286","type":"journal-article","created":{"date-parts":[[2025,3,4]],"date-time":"2025-03-04T12:10:14Z","timestamp":1741090214000},"page":"1-30","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":17,"title":["SocialMind: LLM-based Proactive AR Social Assistive System with Human-like Perception for In-situ Live Interactions"],"prefix":"10.1145","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0032-2539","authenticated-orcid":false,"given":"Bufang","family":"Yang","sequence":"first","affiliation":[{"name":"The Chinese University of Hong Kong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1852-3825","authenticated-orcid":false,"given":"Yunqi","family":"Guo","sequence":"additional","affiliation":[{"name":"The Chinese University of Hong Kong, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-5203-7496","authenticated-orcid":false,"given":"Lilin","family":"Xu","sequence":"additional","affiliation":[{"name":"Columbia University, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4433-5211","authenticated-orcid":false,"given":"Zhenyu","family":"Yan","sequence":"additional","affiliation":[{"name":"The Chinese University of Hong Kong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7206-6584","authenticated-orcid":false,"given":"Hongkai","family":"Chen","sequence":"additional","affiliation":[{"name":"The Chinese University of Hong Kong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1772-7751","authenticated-orcid":false,"given":"Guoliang","family":"Xing","sequence":"additional","affiliation":[{"name":"The Chinese University of Hong Kong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6480-0299","authenticated-orcid":false,"given":"Xiaofan","family":"Jiang","sequence":"additional","affiliation":[{"name":"Columbia University, United States"}]}],"member":"320","published-online":{"date-parts":[[2025,3,4]]},"reference":[{"key":"e_1_2_1_1_1","unstructured":"2024. Apple Siri. https:\/\/www.apple.com\/siri\/."},{"key":"e_1_2_1_2_1","unstructured":"2024. New in Gemini: Gemini Live and connected Google apps in more languages. https:\/\/blog.google\/products\/gemini\/gemini-live-extensions-language-expansion\/."},{"key":"e_1_2_1_3_1","unstructured":"2024. Quality of life indicators - social interactions. https:\/\/ec.europa.eu\/eurostat\/statistics-explained\/index.php?title=Quality_of_life_indicators_-_social_interactions."},{"key":"e_1_2_1_4_1","unstructured":"2024. RayNeo X2. https:\/\/rayneo.cn\/product\/x2\/specs\/."},{"key":"e_1_2_1_5_1","unstructured":"2024. Social Anxiety Disorder. https:\/\/adaa.org\/understanding-anxiety\/social-anxiety-disorder."},{"key":"e_1_2_1_6_1","unstructured":"2024. Tianji. https:\/\/github.com\/SocialAI-tianji."},{"key":"e_1_2_1_7_1","volume-title":"Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al.","author":"Achiam Josh","year":"2023","unstructured":"Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al. 2023. Gpt-4 technical report. arXiv preprint arXiv:2303.08774 (2023)."},{"key":"e_1_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3640794.3665561"},{"key":"e_1_2_1_9_1","volume-title":"distance and affiliation. Sociometry","author":"Argyle Michael","year":"1965","unstructured":"Michael Argyle and Janet Dean. 1965. Eye-contact, distance and affiliation. Sociometry (1965), 289--304."},{"key":"e_1_2_1_10_1","volume-title":"Individual differences and personality","author":"Ashton Michael C","unstructured":"Michael C Ashton. 2022. Individual differences and personality. Academic Press."},{"key":"e_1_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.nlposs-1.24"},{"key":"e_1_2_1_12_1","volume-title":"Mert Yuksekgonul, Jacopo Tagliabue, Dan Jurafsky, and James Zou.","author":"Bianchi Federico","year":"2024","unstructured":"Federico Bianchi, Patrick John Chia, Mert Yuksekgonul, Jacopo Tagliabue, Dan Jurafsky, and James Zou. 2024. How well can llms negotiate? negotiationarena platform and analysis. arXiv preprint arXiv:2402.05863 (2024)."},{"key":"e_1_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jml.2019.104047"},{"key":"e_1_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1080\/10494820902924995"},{"key":"e_1_2_1_15_1","unstructured":"Harrison Chase. 2022. LangChain. https:\/\/github.com\/langchain-ai\/langchain"},{"key":"e_1_2_1_16_1","volume-title":"Listen to Your Face: A Face Authentication Scheme Based on Acoustic Signals. ACM Transactions on Sensor Networks","author":"Chen Huimin","year":"2024","unstructured":"Huimin Chen, Chaojie Gu, Lilin Xu, Rui Tan, Shibo He, and Jiming Chen. 2024. Listen to Your Face: A Face Authentication Scheme Based on Acoustic Signals. ACM Transactions on Sensor Networks (2024)."},{"key":"e_1_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3643832.3661890"},{"key":"e_1_2_1_18_1","volume-title":"Towards Multimodal Emotional Support Conversation Systems. arXiv preprint arXiv:2408.03650","author":"Chu Yuqi","year":"2024","unstructured":"Yuqi Chu, Lizi Liao, Zhiyuan Zhou, Chong-Wah Ngo, and Richang Hong. 2024. Towards Multimodal Emotional Support Conversation Systems. arXiv preprint arXiv:2408.03650 (2024)."},{"key":"e_1_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3626772.3657843"},{"key":"e_1_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDCS.2017.94"},{"key":"e_1_2_1_21_1","volume-title":"Nonverbal communication. Psychological bulletin 72, 2","author":"Starkey Duncan Jr.","year":"1969","unstructured":"Starkey Duncan Jr. 1969. Nonverbal communication. Psychological bulletin 72, 2 (1969), 118."},{"key":"e_1_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3613905.3650764"},{"key":"e_1_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/1631272.1631301"},{"key":"e_1_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1098\/rstb.2009.0142"},{"key":"e_1_2_1_25_1","volume-title":"2024 USENIX Annual Technical Conference (USENIX ATC 24)","author":"Gao Bin","year":"2024","unstructured":"Bin Gao, Zhuomin He, Puru Sharma, Qingxuan Kang, Djordje Jevdjic, Junbo Deng, Xingkun Yang, Zhou Yu, and Pengfei Zuo. 2024. {Cost-Efficient} Large Language Model Serving for Multi-turn Conversations with {CachedAttention}. In 2024 USENIX Annual Technical Conference (USENIX ATC 24). 111--126."},{"key":"e_1_2_1_26_1","volume-title":"Aligning llm agents by learning latent preference from user edits. arXiv preprint arXiv:2404.15269","author":"Gao Ge","year":"2024","unstructured":"Ge Gao, Alexey Taymanov, Eduardo Salinas, Paul Mineiro, and Dipendra Misra. 2024. Aligning llm agents by learning latent preference from user edits. arXiv preprint arXiv:2404.15269 (2024)."},{"key":"e_1_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3678516"},{"key":"e_1_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1007\/3-540-44887-X_98"},{"key":"e_1_2_1_29_1","first-page":"325","article-title":"Prompt cache: Modular attention reuse for low-latency inference","volume":"6","author":"Gim In","year":"2024","unstructured":"In Gim, Guojun Chen, Seung-seob Lee, Nikhil Sarda, Anurag Khandelwal, and Lin Zhong. 2024. Prompt cache: Modular attention reuse for low-latency inference. Proceedings of Machine Learning and Systems 6 (2024), 325--338.","journal-title":"Proceedings of Machine Learning and Systems"},{"key":"e_1_2_1_30_1","unstructured":"Google. 2024. Google Assistant. https:\/\/assistant.google.com\/learn\/."},{"key":"e_1_2_1_31_1","unstructured":"Google. 2024. MediaPipe. https:\/\/github.com\/google\/mediapipe Accessed: 2024-10-30."},{"key":"e_1_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/3241539.3241557"},{"key":"e_1_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3570361.3613260"},{"key":"e_1_2_1_34_1","volume-title":"Nonverbal communication. Annual review of psychology 70, 1","author":"Hall Judith A","year":"2019","unstructured":"Judith A Hall, Terrence G Horgan, and Nora A Murphy. 2019. Nonverbal communication. Annual review of psychology 70, 1 (2019), 271--294."},{"key":"e_1_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.naacl-main.49"},{"key":"e_1_2_1_36_1","volume-title":"Tao Feng, Haolan Zhan, Lizhen Qu, Suraj Sharma, Ingrid Zukerman, Zhaleh Semnani-Azad, et al.","author":"Hua Yuncheng","year":"2024","unstructured":"Yuncheng Hua, Zhuang Li, Linhao Luo, Kadek Ananta Satriadi, Tao Feng, Haolan Zhan, Lizhen Qu, Suraj Sharma, Ingrid Zukerman, Zhaleh Semnani-Azad, et al. 2024. Sadas: A dialogue assistant system towards remediating norm violations in bilingual socio-cultural conversations. arXiv preprint arXiv:2402.01736 (2024)."},{"key":"e_1_2_1_37_1","volume-title":"Assistive Large Language Model Agents for Socially-Aware Negotiation Dialogues. arXiv preprint arXiv:2402.01737","author":"Hua Yuncheng","year":"2024","unstructured":"Yuncheng Hua, Lizhen Qu, and Gholamreza Haffari. 2024. Assistive Large Language Model Agents for Socially-Aware Negotiation Dialogues. arXiv preprint arXiv:2402.01737 (2024)."},{"key":"e_1_2_1_38_1","volume-title":"Apple Vision Pro. https:\/\/www.apple.com\/apple-vision-pro\/ Mixed-reality headset by Apple Inc., announced","author":"Apple Inc. 2024.","year":"2023","unstructured":"Apple Inc. 2024. Apple Vision Pro. https:\/\/www.apple.com\/apple-vision-pro\/ Mixed-reality headset by Apple Inc., announced in 2023, offering advanced augmented and virtual reality experiences."},{"key":"e_1_2_1_39_1","unstructured":"INMO Glass. 2024. INMO Air2 - Next-Gen Wireless AR Glasses. https:\/\/air2.inmoglass.com\/ Accessed: 2024-10-31."},{"key":"e_1_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.findings-acl.904"},{"key":"e_1_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/3613904.3642894"},{"key":"e_1_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01029"},{"key":"e_1_2_1_43_1","volume-title":"Social-LLM: Modeling User Behavior at Scale using Language Models and Social Network Data. arXiv preprint arXiv:2401.00893","author":"Jiang Julie","year":"2023","unstructured":"Julie Jiang and Emilio Ferrara. 2023. Social-LLM: Modeling User Behavior at Scale using Language Models and Social Network Data. arXiv preprint arXiv:2401.00893 (2023)."},{"key":"e_1_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/3678525"},{"key":"e_1_2_1_45_1","volume-title":"Social communication coaching smartglasses: Well tolerated in a diverse sample of children and adults with autism. JMIR mHealth and uHealth 5, 9","author":"Keshav Neha U","year":"2017","unstructured":"Neha U Keshav, Joseph P Salisbury, Arshya Vahabzadeh, and Ned T Sahin. 2017. Social communication coaching smartglasses: Well tolerated in a diverse sample of children and adults with autism. JMIR mHealth and uHealth 5, 9 (2017), e8534."},{"key":"e_1_2_1_46_1","volume-title":"Prosocialdialog: A prosocial backbone for conversational agents. arXiv preprint arXiv:2205.12688","author":"Kim Hyunwoo","year":"2022","unstructured":"Hyunwoo Kim, Youngjae Yu, Liwei Jiang, Ximing Lu, Daniel Khashabi, Gunhee Kim, Yejin Choi, and Maarten Sap. 2022. Prosocialdialog: A prosocial backbone for conversational agents. arXiv preprint arXiv:2205.12688 (2022)."},{"key":"e_1_2_1_47_1","first-page":"9459","article-title":"Retrieval-augmented generation for knowledge-intensive nlp tasks","volume":"33","author":"Lewis Patrick","year":"2020","unstructured":"Patrick Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich K\u00fcttler, Mike Lewis, Wen-tau Yih, Tim Rockt\u00e4schel, et al. 2020. Retrieval-augmented generation for knowledge-intensive nlp tasks. Advances in Neural Information Processing Systems 33 (2020), 9459--9474.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/3384419.3430779"},{"key":"e_1_2_1_49_1","volume-title":"SCALM: Towards Semantic Caching for Automated Chat Services with Large Language Models. arXiv preprint arXiv:2406.00025","author":"Li Jiaxing","year":"2024","unstructured":"Jiaxing Li, Chi Xu, Feng Wang, Isaac M von Riedemann, Cong Zhang, and Jiangchuan Liu. 2024. SCALM: Towards Semantic Caching for Automated Chat Services with Large Language Models. arXiv preprint arXiv:2406.00025 (2024)."},{"key":"e_1_2_1_50_1","volume-title":"Dailydialog: A manually labelled multi-turn dialogue dataset. arXiv preprint arXiv:1710.03957","author":"Li Yanran","year":"2017","unstructured":"Yanran Li, Hui Su, Xiaoyu Shen, Wenjie Li, Ziqiang Cao, and Shuzi Niu. 2017. Dailydialog: A manually labelled multi-turn dialogue dataset. arXiv preprint arXiv:1710.03957 (2017)."},{"key":"e_1_2_1_51_1","volume-title":"Visual instruction tuning. Advances in neural information processing systems 36","author":"Liu Haotian","year":"2024","unstructured":"Haotian Liu, Chunyuan Li, Qingyang Wu, and Yong Jae Lee. 2024. Visual instruction tuning. Advances in neural information processing systems 36 (2024)."},{"key":"e_1_2_1_52_1","volume-title":"Andes: Defining and Enhancing Quality-of-Experience in LLM-Based Text Streaming Services. arXiv preprint arXiv:2404.16283","author":"Liu Jiachen","year":"2024","unstructured":"Jiachen Liu, Zhiyu Wu, Jae-Won Chung, Fan Lai, Myungjin Lee, and Mosharaf Chowdhury. 2024. Andes: Defining and Enhancing Quality-of-Experience in LLM-Based Text Streaming Services. arXiv preprint arXiv:2404.16283 (2024)."},{"key":"e_1_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/3666025.3699428"},{"key":"e_1_2_1_54_1","unstructured":"Chaquopy LLC. 2024. Chaquopy: the Python SDK for Android. https:\/\/github.com\/chaquo\/chaquopy. Accessed: 2024-10-30."},{"key":"e_1_2_1_55_1","volume-title":"Alexandria K Vail, Sunreeta Bhattacharya, \u00c1lvaro Fern\u00e1ndez Garc\u00eda, Kailana Baker-Matsuoka, Sheryl Mathew, Lori L Holt, and Fernando De la Torre.","author":"Ma Cheng Charles","year":"2024","unstructured":"Cheng Charles Ma, Kevin Hyekang Joo, Alexandria K Vail, Sunreeta Bhattacharya, \u00c1lvaro Fern\u00e1ndez Garc\u00eda, Kailana Baker-Matsuoka, Sheryl Mathew, Lori L Holt, and Fernando De la Torre. 2024. Multimodal Fusion with LLMs for Engagement Prediction in Natural Conversation. arXiv preprint arXiv:2409.09135 (2024)."},{"key":"e_1_2_1_56_1","unstructured":"Meta. 2024. Introducing Orion Our First True Augmented Reality Glasses. https:\/\/about.fb.com\/news\/2024\/09\/introducing-orion-our-first-true-augmented-reality-glasses\/ Accessed: 2024-10-31."},{"key":"e_1_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.5334\/joc.268"},{"key":"e_1_2_1_58_1","unstructured":"Microsoft. 2024. Limited Access to Speaker Recognition. https:\/\/learn.microsoft.com\/en-us\/legal\/cognitive-services\/speech-service\/speaker-recognition\/limited-access-speaker-recognition#registration-process."},{"key":"e_1_2_1_59_1","unstructured":"Microsoft Learn. 2024. Speaker Recognition Overview. https:\/\/learn.microsoft.com\/en-us\/azure\/ai-services\/speech-service\/speaker-recognition-overview Accessed: 2024-10-31."},{"key":"e_1_2_1_60_1","volume-title":"Jennifer Neville, and Tara Safavi.","author":"Mysore Sheshera","year":"2023","unstructured":"Sheshera Mysore, Zhuoran Lu, Mengting Wan, Longqi Yang, Steve Menezes, Tina Baghaee, Emmanuel Barajas Gonzalez, Jennifer Neville, and Tara Safavi. 2023. Pearl: Personalizing large language model writing assistants with generation-calibrated retrievers. arXiv preprint arXiv:2311.09180 (2023)."},{"key":"e_1_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.5555\/1953048.2078195"},{"key":"e_1_2_1_62_1","unstructured":"Even Realities. 2024. G1: Next-Gen Smart Glasses with Display. https:\/\/www.evenrealities.com\/g1 Accessed: 2024-10-31."},{"key":"e_1_2_1_63_1","volume-title":"Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. arXiv preprint arXiv:1908.10084","author":"Reimers N","year":"2019","unstructured":"N Reimers. 2019. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. arXiv preprint arXiv:1908.10084 (2019)."},{"key":"e_1_2_1_64_1","volume-title":"Speech ReaLLM--Real-time Streaming Speech Recognition with Multimodal LLMs by Teaching the Flow of Time. arXiv preprint arXiv:2406.09569","author":"Seide Frank","year":"2024","unstructured":"Frank Seide, Morrie Doulaty, Yangyang Shi, Yashesh Gaur, Junteng Jia, and Chunyang Wu. 2024. Speech ReaLLM--Real-time Streaming Speech Recognition with Multimodal LLMs by Teaching the Flow of Time. arXiv preprint arXiv:2406.09569 (2024)."},{"key":"e_1_2_1_65_1","unstructured":"WIRED Staff. 2024. XRAI Glass Caption AR Glasses: First Look. https:\/\/www.wired.com\/story\/xrai-glass-caption-ar-glasses-first-look\/ Accessed: 2024-10-31."},{"key":"e_1_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.2116915119"},{"key":"e_1_2_1_67_1","unstructured":"Hugo Touvron Louis Martin Kevin Stone Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale et al. 2023. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)."},{"key":"e_1_2_1_68_1","doi-asserted-by":"publisher","DOI":"10.1145\/3699751"},{"key":"e_1_2_1_69_1","volume-title":"Denny Zhou, et al.","author":"Wei Jason","year":"2022","unstructured":"Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Fei Xia, Ed Chi, Quoc V Le, Denny Zhou, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in neural information processing systems 35 (2022), 24824--24837."},{"key":"e_1_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2018.07.073"},{"key":"e_1_2_1_71_1","volume-title":"Talk While Thinking in Streaming. arXiv preprint arXiv:2408.16725","author":"Xie Zhifei","year":"2024","unstructured":"Zhifei Xie and Changqiao Wu. 2024. Mini-Omni: Language Models Can Hear, Talk While Thinking in Streaming. arXiv preprint arXiv:2408.16725 (2024)."},{"key":"e_1_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDCS60910.2024.00103"},{"key":"e_1_2_1_73_1","doi-asserted-by":"publisher","DOI":"10.1145\/3241539.3241563"},{"key":"e_1_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1186\/s41070-016-0011-8"},{"key":"e_1_2_1_75_1","first-page":"1","article-title":"Can Large Language Models Be Good Companions? An LLM-Based Eyewear System with Conversational Common Ground","volume":"8","author":"Xu Zhenyu","year":"2024","unstructured":"Zhenyu Xu, Hailin Xu, Zhouyang Lu, Yingying Zhao, Rui Zhu, Yujiang Wang, Mingzhi Dong, Yuhu Chang, Qin Lv, Robert P Dick, et al. 2024. Can Large Language Models Be Good Companions? An LLM-Based Eyewear System with Conversational Common Ground. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 8, 2 (2024), 1--41.","journal-title":"Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies"},{"key":"e_1_2_1_76_1","doi-asserted-by":"publisher","DOI":"10.1145\/3625687.3625793"},{"key":"e_1_2_1_77_1","volume-title":"VIAssist: Adapting Multi-modal Large Language Models for Users with Visual Impairments. arXiv preprint arXiv:2404.02508","author":"Yang Bufang","year":"2024","unstructured":"Bufang Yang, Lixing He, Kaiwei Liu, and Zhenyu Yan. 2024. VIAssist: Adapting Multi-modal Large Language Models for Users with Visual Impairments. arXiv preprint arXiv:2404.02508 (2024)."},{"key":"e_1_2_1_78_1","doi-asserted-by":"publisher","DOI":"10.1145\/3699765"},{"key":"e_1_2_1_79_1","doi-asserted-by":"publisher","DOI":"10.1145\/3699765"},{"key":"e_1_2_1_80_1","volume-title":"SirLLM: Streaming infinite retentive LLM. arXiv preprint arXiv:2405.12528","author":"Yao Yao","year":"2024","unstructured":"Yao Yao, Zuchao Li, and Hai Zhao. 2024. SirLLM: Streaming infinite retentive LLM. arXiv preprint arXiv:2405.12528 (2024)."},{"key":"e_1_2_1_81_1","volume-title":"Kelly Rosalin, Jureynolds Jureynolds, et al.","author":"Zhan Haolan","year":"2024","unstructured":"Haolan Zhan, Zhuang Li, Xiaoxi Kang, Tao Feng, Yuncheng Hua, Lizhen Qu, Yi Ying, Mei Rianto Chandra, Kelly Rosalin, Jureynolds Jureynolds, et al. 2024. RENOVI: A Benchmark Towards Remediating Norm Violations in Socio-Cultural Conversations. In Findings of the Association for Computational Linguistics: NAACL 2024. 3104--3117."},{"key":"e_1_2_1_82_1","doi-asserted-by":"publisher","DOI":"10.1145\/3539618.3591877"},{"key":"e_1_2_1_83_1","volume-title":"Findings of the Association for Computational Linguistics: EACL 2024. 2019","author":"Zhan Haolan","year":"2024","unstructured":"Haolan Zhan, Yufei Wang, Zhuang Li, Tao Feng, Yuncheng Hua, Suraj Sharma, Lizhen Qu, Zhaleh Semnani Azad, Ingrid Zukerman, and Reza Haf. 2024. Let's Negotiate! A Survey of Negotiation Dialogue Systems. In Findings of the Association for Computational Linguistics: EACL 2024. 2019--2031."},{"key":"e_1_2_1_84_1","first-page":"46595","article-title":"Judging llm-as-a-judge with mt-bench and chatbot arena","volume":"36","author":"Zheng Lianmin","year":"2023","unstructured":"Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, Zi Lin, Zhuohan Li, Dacheng Li, Eric Xing, et al. 2023. Judging llm-as-a-judge with mt-bench and chatbot arena. Advances in Neural Information Processing Systems 36 (2023), 46595--46623.","journal-title":"Advances in Neural Information Processing Systems"}],"container-title":["Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3712286","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3712286","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T19:31:11Z","timestamp":1755891071000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3712286"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,3,3]]},"references-count":84,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,3,3]]}},"alternative-id":["10.1145\/3712286"],"URL":"https:\/\/doi.org\/10.1145\/3712286","relation":{},"ISSN":["2474-9567"],"issn-type":[{"value":"2474-9567","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,3,3]]},"assertion":[{"value":"2025-03-04","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}