{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,13]],"date-time":"2026-02-13T06:09:59Z","timestamp":1770962999622,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":72,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,5,6]],"date-time":"2025-05-06T00:00:00Z","timestamp":1746489600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62202407"],"award-info":[{"award-number":["62202407"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002920","name":"Research Grants Council, University Grants Committee","doi-asserted-by":"publisher","award":["GRF 14214022"],"award-info":[{"award-number":["GRF 14214022"]}],"id":[{"id":"10.13039\/501100002920","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002920","name":"Research Grants Council, University Grants Committee","doi-asserted-by":"publisher","award":["GRF 14212323"],"award-info":[{"award-number":["GRF 14212323"]}],"id":[{"id":"10.13039\/501100002920","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002920","name":"Research Grants Council, University Grants Committee","doi-asserted-by":"publisher","award":["TRS T43-513\/23-N"],"award-info":[{"award-number":["TRS T43-513\/23-N"]}],"id":[{"id":"10.13039\/501100002920","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,5,6]]},"DOI":"10.1145\/3715014.3722070","type":"proceedings-article","created":{"date-parts":[[2025,5,4]],"date-time":"2025-05-04T23:37:21Z","timestamp":1746401841000},"page":"213-225","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":4,"title":["TaskSense: A Translation-like Approach for Tasking Heterogeneous Sensor Systems with LLMs"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0002-4108-0898","authenticated-orcid":false,"given":"Kaiwei","family":"Liu","sequence":"first","affiliation":[{"name":"The Chinese University of Hong Kong, Hong Kong, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0032-2539","authenticated-orcid":false,"given":"Bufang","family":"Yang","sequence":"additional","affiliation":[{"name":"The Chinese University of Hong Kong, Hong kong, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-5203-7496","authenticated-orcid":false,"given":"Lilin","family":"Xu","sequence":"additional","affiliation":[{"name":"The Chinese University of Hong Kong, Hong Kong, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1852-3825","authenticated-orcid":false,"given":"Yunqi","family":"Guo","sequence":"additional","affiliation":[{"name":"The Chinese University of Hong Kong, Hong Kong, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1772-7751","authenticated-orcid":false,"given":"Guoliang","family":"Xing","sequence":"additional","affiliation":[{"name":"The Chinese University of Hong Kong, Hong Kong, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6750-6706","authenticated-orcid":false,"given":"Xian","family":"Shuai","sequence":"additional","affiliation":[{"name":"Noah's Ark Lab, Huawei Technologies, Hong Kong, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0432-5510","authenticated-orcid":false,"given":"Xiaozhe","family":"Ren","sequence":"additional","affiliation":[{"name":"Noah's Ark Lab, Huawei Technologies, Hong Kong, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9117-8247","authenticated-orcid":false,"given":"Xin","family":"Jiang","sequence":"additional","affiliation":[{"name":"Noah's Ark Lab, Huawei Technologies, Hong Kong, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4433-5211","authenticated-orcid":false,"given":"Zhenyu","family":"Yan","sequence":"additional","affiliation":[{"name":"The Chinese University of Hong Kong, Hong Kong, Hong Kong"}]}],"member":"320","published-online":{"date-parts":[[2025,5,6]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"2023. Mistral Inference. https:\/\/github.com\/mistralai\/mistral-inference ."},{"key":"e_1_3_2_1_2_1","unstructured":"2024. Autonomous & Sensor Technology. https:\/\/www.statista.com\/outlook\/tmo\/artificial-intelligence\/autonomous-sensor-technology\/worldwide ."},{"key":"e_1_3_2_1_3_1","unstructured":"2024. Hello GPT-4o. https:\/\/openai.com\/index\/hello-gpt-4o\/ ."},{"key":"e_1_3_2_1_4_1","unstructured":"2024. International Labour Organization. https:\/\/www.ilo.org\/ ."},{"key":"e_1_3_2_1_5_1","volume-title":"Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al.","author":"Achiam Josh","year":"2023","unstructured":"Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al. 2023. Gpt-4 technical report. arXiv preprint arXiv:2303.08774 (2023)."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/800125.804056"},{"key":"e_1_3_2_1_7_1","unstructured":"Cohere AI. 2023. Introduction to Text Embeddings. https:\/\/github.com\/cohere-ai\/notebooks\/blob\/main\/notebooks\/llmu\/Introduction_Text_Embeddings.ipynb?ref=cohere-ai.ghost.io."},{"key":"e_1_3_2_1_8_1","volume-title":"Amazon Bedrock: Build Generative AI Applications with Foundation Models. https:\/\/aws.amazon.com\/bedrock\/. Accessed: 2024-06-30.","author":"Services Amazon Web","year":"2024","unstructured":"Amazon Web Services. 2024. Amazon Bedrock: Build Generative AI Applications with Foundation Models. https:\/\/aws.amazon.com\/bedrock\/. Accessed: 2024-06-30."},{"key":"e_1_3_2_1_9_1","volume-title":"Home Assistance: Awaken Your Home. https:\/\/www.home-assistant.io\/.","author":"Assistance Home","year":"2024","unstructured":"Home Assistance. 2024. Home Assistance: Awaken Your Home. https:\/\/www.home-assistant.io\/."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.5664\/jcsm.27351"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3708324"},{"key":"e_1_3_2_1_12_1","volume-title":"Maybe only 0.5% data is needed: A preliminary exploration of low training data instruction tuning. arXiv preprint arXiv:2305.09246","author":"Chen Hao","year":"2023","unstructured":"Hao Chen, Yiming Zhang, Qi Zhang, Hantao Yang, Xiaomeng Hu, Xuetao Ma, Yifan Yanggong, and Junbo Zhao. 2023. Maybe only 0.5% data is needed: A preliminary exploration of low training data instruction tuning. arXiv preprint arXiv:2305.09246 (2023)."},{"key":"e_1_3_2_1_13_1","first-page":"492","article-title":"International physical activity questionnaire-short form","volume":"65","author":"Craig C","year":"2017","unstructured":"C Craig, A Marshall, M Sjostrom, A Bauman, P Lee, D Macfarlane, T Lam, and S Stewart. 2017. International physical activity questionnaire-short form. J Am Coll Health 65, 7 (2017), 492--501.","journal-title":"J Am Coll Health"},{"key":"e_1_3_2_1_14_1","volume-title":"Llmind: Orchestrating ai and iot with llms for complex task execution. arXiv preprint arXiv:2312.09007","author":"Cui Hongwei","year":"2023","unstructured":"Hongwei Cui, Yuyang Du, Qun Yang, Yulin Shao, and Soung Chang Liew. 2023. Llmind: Orchestrating ai and iot with llms for complex task execution. arXiv preprint arXiv:2312.09007 (2023)."},{"key":"e_1_3_2_1_15_1","volume-title":"What Linguistic Features and Languages are Important in LLM Translation? arXiv preprint arXiv:2402.13917","author":"Diandaru Ryandito","year":"2024","unstructured":"Ryandito Diandaru, Lucky Susanto, Zilu Tang, Ayu Purwarianti, and Derry Wijaya. 2024. What Linguistic Features and Languages are Important in LLM Translation? arXiv preprint arXiv:2402.13917 (2024)."},{"key":"e_1_3_2_1_16_1","unstructured":"Hugging Face. 2024. FFmpeg: A complete cross-platform solution to record convert and stream audio and video. https:\/\/www.ffmpeg.org\/."},{"key":"e_1_3_2_1_17_1","unstructured":"Hugging Face. 2024. Hugging Face Community. https:\/\/huggingface.co\/."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijporl.2019.109733"},{"key":"e_1_3_2_1_19_1","volume-title":"Armand Joulin, and Ishan Misra.","author":"Girdhar Rohit","year":"2023","unstructured":"Rohit Girdhar, Alaaeldin El-Nouby, Zhuang Liu, Mannat Singh, Kalyan Vasudev Alwala, Armand Joulin, and Ishan Misra. 2023. ImageBind: One Embedding Space To Bind Them All. In CVPR."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"crossref","unstructured":"Christopher G Goetz Barbara C Tilley Stephanie R Shaftman Glenn T Stebbins Stanley Fahn Pablo Martinez-Martin Werner Poewe Cristina Sampaio Matthew B Stern Richard Dodel et al. 2008. Movement Disorder Society-sponsored revision of the Unified Parkinson's Disease Rating Scale (MDS-UPDRS): scale presentation and clinimetric testing results. Movement disorders: official journal of the Movement Disorder Society 23 15 (2008) 2129--2170.","DOI":"10.1002\/mds.22340"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.02510"},{"key":"e_1_3_2_1_22_1","volume-title":"GPT-4, Claude-3, and Mistral-Large. arXiv preprint arXiv:2405.05444","author":"Jauhiainen Jussi S","year":"2024","unstructured":"Jussi S Jauhiainen and Agust\u00edn Garagorry Guerra. 2024. Evaluating Students' Open-ended Written Responses with LLMs: Using the RAG Framework for GPT-3.5, GPT-4, Claude-3, and Mistral-Large. arXiv preprint arXiv:2405.05444 (2024)."},{"key":"e_1_3_2_1_23_1","volume-title":"Sung Ju Hwang, and Jong C Park","author":"Jeong Soyeong","year":"2024","unstructured":"Soyeong Jeong, Jinheon Baek, Sukmin Cho, Sung Ju Hwang, and Jong C Park. 2024. Database-Augmented Query Representation for Information Retrieval. arXiv preprint arXiv:2406.16013 (2024)."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00065"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3643505"},{"key":"e_1_3_2_1_26_1","volume-title":"Statistical machine translation","author":"Koehn Philipp","unstructured":"Philipp Koehn. 2009. Statistical machine translation. Cambridge University Press."},{"key":"e_1_3_2_1_27_1","volume-title":"Machel Reid, Yutaka Matsuo, and Yusuke Iwasawa.","author":"Kojima Takeshi","year":"2022","unstructured":"Takeshi Kojima, Shixiang Shane Gu, Machel Reid, Yutaka Matsuo, and Yusuke Iwasawa. 2022. Large language models are zero-shot reasoners. Advances in neural information processing systems 35 (2022), 22199--22213."},{"key":"e_1_3_2_1_28_1","volume-title":"Transllama: Llm-based simultaneous translation system. arXiv preprint arXiv:2402.04636","author":"Koshkin Roman","year":"2024","unstructured":"Roman Koshkin, Katsuhito Sudoh, and Satoshi Nakamura. 2024. Transllama: Llm-based simultaneous translation system. arXiv preprint arXiv:2402.04636 (2024)."},{"key":"e_1_3_2_1_29_1","unstructured":"Labelbox. 2024. GPT-3.5: Models - OpenAI. https:\/\/labelbox.com\/product\/model\/foundry-models\/gpt-3-5\/. Accessed: 2024-06-30."},{"key":"e_1_3_2_1_30_1","volume-title":"Cross-lingual language model pretraining. arXiv preprint arXiv:1901.07291","author":"Lample Guillaume","year":"2019","unstructured":"Guillaume Lample and Alexis Conneau. 2019. Cross-lingual language model pretraining. arXiv preprint arXiv:1901.07291 (2019)."},{"key":"e_1_3_2_1_31_1","first-page":"2","article-title":"Brody instrumental activities of daily living scale (IADL)","volume":"108","author":"Lawton McMahon E","year":"2008","unstructured":"McMahon E Lawton. 2008. Brody instrumental activities of daily living scale (IADL). MaineHealth 108, 4 (2008), 2.","journal-title":"MaineHealth"},{"key":"e_1_3_2_1_32_1","volume-title":"Self-alignment with instruction backtranslation. arXiv preprint arXiv:2308.06259","author":"Li Xian","year":"2023","unstructured":"Xian Li, Ping Yu, Chunting Zhou, Timo Schick, Luke Zettlemoyer, Omer Levy, Jason Weston, and Mike Lewis. 2023. Self-alignment with instruction backtranslation. arXiv preprint arXiv:2308.06259 (2023)."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"crossref","unstructured":"Yaobo Liang Chenfei Wu Ting Song Wenshan Wu Yan Xia Yu Liu Yang Ou Shuai Lu Lei Ji Shaoguang Mao et al. 2023. Taskmatrix. ai: Completing tasks by connecting foundation models with millions of apis. arXiv preprint arXiv:2303.16434 (2023).","DOI":"10.34133\/icomputing.0063"},{"key":"e_1_3_2_1_34_1","unstructured":"Haotian Liu Chunyuan Li Qingyang Wu and Yong Jae Lee. 2023. Visual Instruction Tuning. In NeurIPS."},{"key":"e_1_3_2_1_35_1","volume-title":"Controlllm: Augment language models with tools by searching on graphs. arXiv preprint arXiv:2310.17796","author":"Liu Zhaoyang","year":"2023","unstructured":"Zhaoyang Liu, Zeqiang Lai, Zhangwei Gao, Erfei Cui, Zhiheng Li, Xizhou Zhu, Lewei Lu, Qifeng Chen, Yu Qiao, Jifeng Dai, et al. 2023. Controlllm: Augment language models with tools by searching on graphs. arXiv preprint arXiv:2310.17796 (2023)."},{"key":"e_1_3_2_1_36_1","volume-title":"Yee Whye Teh, and Tom Rainforth","author":"Miao Ning","year":"2023","unstructured":"Ning Miao, Yee Whye Teh, and Tom Rainforth. 2023. Selfcheck: Using llms to zero-shot check their own step-by-step reasoning. arXiv preprint arXiv:2308.00436 (2023)."},{"key":"e_1_3_2_1_37_1","unstructured":"Microsoft. 2024. Azure OpenAI Service. https:\/\/azure.microsoft.com\/en-us\/products\/ai-services\/openai-service."},{"key":"e_1_3_2_1_38_1","unstructured":"NVIDIA Corporation. 2023. Jetson Orin Modules and Developer Kits. https:\/\/www.nvidia.com\/en-us\/autonomous-machines\/embedded-systems\/jetson-orin\/"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3636534.3649370"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"crossref","unstructured":"Xiaomin Ouyang Xian Shuai Yang Li Li Pan Xifan Zhang Heming Fu Xinyan Wang Shihua Cao Jiang Xin Hazel Mok et al. 2023. ADMarker: A Multi-Modal Federated Learning System for Monitoring Digital Biomarkers of Alzheimer's Disease. arXiv preprint arXiv:2310.15301 (2023).","DOI":"10.1145\/3636534.3649370"},{"key":"e_1_3_2_1_41_1","volume-title":"LLMSense: Harnessing LLMs for High-level Reasoning Over Spatiotemporal Sensor Traces. arXiv preprint arXiv:2403.19857","author":"Ouyang Xiaomin","year":"2024","unstructured":"Xiaomin Ouyang and Mani Srivastava. 2024. LLMSense: Harnessing LLMs for High-level Reasoning Over Spatiotemporal Sensor Traces. arXiv preprint arXiv:2403.19857 (2024)."},{"key":"e_1_3_2_1_42_1","volume-title":"Kwaiagents: Generalized information-seeking agent system with large language models. arXiv preprint arXiv:2312.04889","author":"Pan Haojie","year":"2023","unstructured":"Haojie Pan, Zepeng Zhai, Hao Yuan, Yaojia Lv, Ruiji Fu, Ming Liu, Zhongyuan Wang, and Bing Qin. 2023. Kwaiagents: Generalized information-seeking agent system with large language models. arXiv preprint arXiv:2312.04889 (2023)."},{"key":"e_1_3_2_1_43_1","unstructured":"Poe. 2024. Poe Platform. https:\/\/poe.com\/."},{"key":"e_1_3_2_1_44_1","volume-title":"Toolllm: Facilitating large language models to master 16000+ real-world apis. arXiv preprint arXiv:2307.16789","author":"Qin Yujia","year":"2023","unstructured":"Yujia Qin, Shihao Liang, Yining Ye, Kunlun Zhu, Lan Yan, Yaxi Lu, Yankai Lin, Xin Cong, Xiangru Tang, Bill Qian, et al. 2023. Toolllm: Facilitating large language models to master 16000+ real-world apis. arXiv preprint arXiv:2307.16789 (2023)."},{"key":"e_1_3_2_1_45_1","volume-title":"Toolformer: Language models can teach themselves to use tools. Advances in Neural Information Processing Systems 36","author":"Schick Timo","year":"2024","unstructured":"Timo Schick, Jane Dwivedi-Yu, Roberto Dess\u00ec, Roberta Raileanu, Maria Lomeli, Eric Hambro, Luke Zettlemoyer, Nicola Cancedda, and Thomas Scialom. 2024. Toolformer: Language models can teach themselves to use tools. Advances in Neural Information Processing Systems 36 (2024)."},{"key":"e_1_3_2_1_46_1","volume-title":"Conference on robot learning. PMLR, 492--504","author":"Shah Dhruv","year":"2023","unstructured":"Dhruv Shah, B\u0142a\u017cej Osi\u0144ski, Sergey Levine, et al. 2023. Lm-nav: Robotic navigation with large pre-trained models of language, vision, and action. In Conference on robot learning. PMLR, 492--504."},{"key":"e_1_3_2_1_47_1","volume-title":"STOP, THAT and one hundred other sleep scales","author":"Shahid Azmeh","unstructured":"Azmeh Shahid, Kate Wilkinson, Shai Marcu, and Colin M Shapiro. 2011. Fatigue assessment scale (FAS). In STOP, THAT and one hundred other sleep scales. Springer, 161--162."},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/MCOM.001.2300550"},{"key":"e_1_3_2_1_49_1","volume-title":"Hugginggpt: Solving ai tasks with chatgpt and its friends in hugging face. Advances in Neural Information Processing Systems 36","author":"Shen Yongliang","year":"2024","unstructured":"Yongliang Shen, Kaitao Song, Xu Tan, Dongsheng Li, Weiming Lu, and Yueting Zhuang. 2024. Hugginggpt: Solving ai tasks with chatgpt and its friends in hugging face. Advances in Neural Information Processing Systems 36 (2024)."},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/3636534.3649352"},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543513"},{"key":"e_1_3_2_1_52_1","unstructured":"Spotify. 2024. Approximate Nearest Neighbors Oh Yeah. https:\/\/github.com\/spotify\/annoy."},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1080\/13607863.2015.1047323"},{"key":"e_1_3_2_1_54_1","volume-title":"Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timoth\u00e9e Lacroix, Baptiste Rozi\u00e8re, Naman Goyal, Eric Hambro, Faisal Azhar, et al. 2023. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)."},{"key":"e_1_3_2_1_55_1","unstructured":"TP-link. 2024. TP-link Smart Home Community. https:\/\/community.tp-link.com\/en\/smart-home\/."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2017.67"},{"key":"e_1_3_2_1_57_1","unstructured":"Vzense. 2024. Vzense Technology. https:\/\/www.vzense.com\/."},{"key":"e_1_3_2_1_58_1","volume-title":"Voyager: An open-ended embodied agent with large language models. arXiv preprint arXiv:2305.16291","author":"Wang Guanzhi","year":"2023","unstructured":"Guanzhi Wang, Yuqi Xie, Yunfan Jiang, Ajay Mandlekar, Chaowei Xiao, Yuke Zhu, Linxi Fan, and Anima Anandkumar. 2023. Voyager: An open-ended embodied agent with large language models. arXiv preprint arXiv:2305.16291 (2023)."},{"key":"e_1_3_2_1_59_1","volume-title":"Wizardlm: Empowering large language models to follow complex instructions. arXiv preprint arXiv:2304.12244","author":"Xu Can","year":"2023","unstructured":"Can Xu, Qingfeng Sun, Kai Zheng, Xiubo Geng, Pu Zhao, Jiazhan Feng, Chongyang Tao, and Daxin Jiang. 2023. Wizardlm: Empowering large language models to follow complex instructions. arXiv preprint arXiv:2304.12244 (2023)."},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/3625687.3625782"},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDCS60910.2024.00103"},{"key":"e_1_3_2_1_62_1","volume-title":"SocialMind: LLM-based Proactive AR Social Assistive System with Human-like Perception for In-situ Live Interactions. arXiv preprint arXiv:2412.04036","author":"Yang Bufang","year":"2024","unstructured":"Bufang Yang, Yunqi Guo, Lilin Xu, Zhenyu Yan, Hongkai Chen, Guoliang Xing, and Xiaofan Jiang. 2024. SocialMind: LLM-based Proactive AR Social Assistive System with Human-like Perception for In-situ Live Interactions. arXiv preprint arXiv:2412.04036 (2024)."},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1145\/3625687.3625793"},{"key":"e_1_3_2_1_64_1","volume-title":"VIAssist: Adapting Multi-modal Large Language Models for Users with Visual Impairments. arXiv preprint arXiv:2404.02508","author":"Yang Bufang","year":"2024","unstructured":"Bufang Yang, Lixing He, Kaiwei Liu, and Zhenyu Yan. 2024. VIAssist: Adapting Multi-modal Large Language Models for Users with Visual Impairments. arXiv preprint arXiv:2404.02508 (2024)."},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1145\/3699765"},{"key":"e_1_3_2_1_66_1","volume-title":"BrainZ-BP: A Non-invasive Cuff-less Blood Pressure Estimation Approach Leveraging Brain Bio-impedance and Electrocardiogram","author":"Yang Bufang","year":"2023","unstructured":"Bufang Yang, Le Liu, Wenxuan Wu, Mengliang Zhou, Hongxing Liu, and Xinbao Ning. 2023. BrainZ-BP: A Non-invasive Cuff-less Blood Pressure Estimation Approach Leveraging Brain Bio-impedance and Electrocardiogram. IEEE Transactions on Instrumentation and Measurement (2023)."},{"key":"e_1_3_2_1_67_1","volume-title":"Are You Being Tracked? Discover the Power of Zero-Shot Trajectory Tracing with LLMs! arXiv preprint arXiv:2403.06201","author":"Yang Huanqi","year":"2024","unstructured":"Huanqi Yang, Sijie Ji, Rucheng Wu, and Weitao Xu. 2024. Are You Being Tracked? Discover the Power of Zero-Shot Trajectory Tracing with LLMs! arXiv preprint arXiv:2403.06201 (2024)."},{"key":"e_1_3_2_1_68_1","volume-title":"React: Synergizing reasoning and acting in language models. arXiv preprint arXiv:2210.03629","author":"Yao Shunyu","year":"2022","unstructured":"Shunyu Yao, Jeffrey Zhao, Dian Yu, Nan Du, Izhak Shafran, Karthik Narasimhan, and Yuan Cao. 2022. React: Synergizing reasoning and acting in language models. arXiv preprint arXiv:2210.03629 (2022)."},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.1145\/3636534.3649361"},{"key":"e_1_3_2_1_70_1","volume-title":"Language agent tree search unifies reasoning acting and planning in language models. arXiv preprint arXiv:2310.04406","author":"Zhou Andy","year":"2023","unstructured":"Andy Zhou, Kai Yan, Michal Shlapentokh-Rothman, Haohan Wang, and Yu-Xiong Wang. 2023. Language agent tree search unifies reasoning acting and planning in language models. arXiv preprint arXiv:2310.04406 (2023)."},{"key":"e_1_3_2_1_71_1","volume-title":"Lima: Less is more for alignment. Advances in Neural Information Processing Systems 36","author":"Zhou Chunting","year":"2024","unstructured":"Chunting Zhou, Pengfei Liu, Puxin Xu, Srinivasan Iyer, Jiao Sun, Yuning Mao, Xuezhe Ma, Avia Efrat, Ping Yu, Lili Yu, et al. 2024. Lima: Less is more for alignment. Advances in Neural Information Processing Systems 36 (2024)."},{"key":"e_1_3_2_1_72_1","volume-title":"Large language models for information retrieval: A survey. arXiv preprint arXiv:2308.07107","author":"Zhu Yutao","year":"2023","unstructured":"Yutao Zhu, Huaying Yuan, Shuting Wang, Jiongnan Liu, Wenhan Liu, Chenlong Deng, Zhicheng Dou, and Ji-Rong Wen. 2023. Large language models for information retrieval: A survey. arXiv preprint arXiv:2308.07107 (2023)."}],"event":{"name":"SenSys '25: 23rd ACM Conference on Embedded Networked Sensor Systems","location":"UC Irvine Student Center. Irvine CA USA","acronym":"SenSys '25","sponsor":["SIGARCH ACM Special Interest Group on Computer Architecture","SIGMETRICS ACM Special Interest Group on Measurement and Evaluation","SIGOPS ACM Special Interest Group on Operating Systems","SIGMOBILE ACM Special Interest Group on Mobility of Systems, Users, Data and Computing","SIGBED ACM Special Interest Group on Embedded Systems"]},"container-title":["Proceedings of the 23rd ACM Conference on Embedded Networked Sensor Systems"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3715014.3722070","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:56:51Z","timestamp":1750298211000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3715014.3722070"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,6]]},"references-count":72,"alternative-id":["10.1145\/3715014.3722070","10.1145\/3715014"],"URL":"https:\/\/doi.org\/10.1145\/3715014.3722070","relation":{},"subject":[],"published":{"date-parts":[[2025,5,6]]},"assertion":[{"value":"2025-05-06","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}