{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,12]],"date-time":"2025-12-12T00:06:53Z","timestamp":1765498013195,"version":"3.48.0"},"publisher-location":"New York, NY, USA","reference-count":16,"publisher":"ACM","funder":[{"DOI":"10.13039\/501100013129","name":"Ministry of SMEs and Startups","doi-asserted-by":"publisher","award":["20277181"],"award-info":[{"award-number":["20277181"]}],"id":[{"id":"10.13039\/501100013129","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,11,10]]},"DOI":"10.1145\/3746252.3761477","type":"proceedings-article","created":{"date-parts":[[2025,11,7]],"date-time":"2025-11-07T23:55:33Z","timestamp":1762559733000},"page":"6659-6663","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["EdgeSLU: 1.58-bit Voice Control Framework"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-3208-4882","authenticated-orcid":false,"given":"Seungeon","family":"Lee","sequence":"first","affiliation":[{"name":"ENERZAi, Seoul, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5572-0713","authenticated-orcid":false,"given":"Junuk","family":"Jung","sequence":"additional","affiliation":[{"name":"ENERZAi, Seoul, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-9395-1821","authenticated-orcid":false,"given":"Sanghyun","family":"Jung","sequence":"additional","affiliation":[{"name":"ENERZAi, Seoul, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6233-0908","authenticated-orcid":false,"given":"Changbeom","family":"Kang","sequence":"additional","affiliation":[{"name":"ENERZAi, Seoul, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-8588-6483","authenticated-orcid":false,"given":"Jaeyoon","family":"Yoo","sequence":"additional","affiliation":[{"name":"ENERZAi, Seoul, Republic of Korea"}]}],"member":"320","published-online":{"date-parts":[[2025,11,10]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2019.2921977"},{"key":"e_1_3_2_1_2_1","volume-title":"Bert for joint intent classification and slot filling. arXiv preprint arXiv:1902.10909","author":"Chen Qian","year":"2019","unstructured":"Qian Chen, Zhu Zhuo, and Wen Wang. 2019. Bert for joint intent classification and slot filling. arXiv preprint arXiv:1902.10909 (2019)."},{"key":"e_1_3_2_1_3_1","volume-title":"FLEURS: Few-shot Learning Evaluation of Universal Representations of Speech. arXiv preprint arXiv:2205.12446","author":"Conneau Alexis","year":"2022","unstructured":"Alexis Conneau, Min Ma, Simran Khanuja, Yu Zhang, Vera Axelrod, Siddharth Dalmia, Jason Riesa, Clara Rivera, and Ankur Bapna. 2022. FLEURS: Few-shot Learning Evaluation of Universal Representations of Speech. arXiv preprint arXiv:2205.12446 (2022). https:\/\/arxiv.org\/abs\/2205.12446"},{"key":"e_1_3_2_1_4_1","unstructured":"Alice Coucke Alaa Saade Adrien Ball Th\u00e9odore Bluche Alexandre Caulier David Leroy Cl\u00e9ment Doumouro Thibault Gisselbrecht Francesco Caltagirone Thibaut Lavril et al. 2018. Snips voice platform: an embedded spoken language understanding system for private-by-design voice interfaces. arXiv preprint arXiv:1805.10190 (2018)."},{"key":"e_1_3_2_1_5_1","volume-title":"Distil-whisper: Robust knowledge distillation via large-scale pseudo labelling. arXiv preprint arXiv:2311.00430","author":"Gandhi Sanchit","year":"2023","unstructured":"Sanchit Gandhi, Patrick von Platen, and Alexander M Rush. 2023. Distil-whisper: Robust knowledge distillation via large-scale pseudo labelling. arXiv preprint arXiv:2311.00430 (2023)."},{"key":"e_1_3_2_1_6_1","volume-title":"2003 IEEE Workshop on Automatic Speech Recognition and Understanding (IEEE Cat. No. 03EX721)","author":"He Yulan","year":"2003","unstructured":"Yulan He and Steve Young. 2003. A data-driven spoken language understanding system. In 2003 IEEE Workshop on Automatic Speech Recognition and Understanding (IEEE Cat. No. 03EX721). IEEE, 583--588."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCRD54409.2022.9730411"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN60899.2024.10650109"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.3390\/computers12030060"},{"key":"e_1_3_2_1_10_1","volume-title":"The era of 1- bit llms: All large language models are in 1.58 bits. arXiv preprint arXiv:2402.17764 1","author":"Ma Shuming","year":"2024","unstructured":"Shuming Ma, Hongyu Wang, Lingxiao Ma, Lei Wang, Wenhui Wang, Shaohan Huang, Lifeng Dong, Ruiping Wang, Jilong Xue, and Furu Wei. 2024. The era of 1- bit llms: All large language models are in 1.58 bits. arXiv preprint arXiv:2402.17764 1 (2024)."},{"key":"e_1_3_2_1_11_1","volume-title":"International conference on machine learning. PMLR, 28492--28518","author":"Radford Alec","year":"2023","unstructured":"Alec Radford, JongWook Kim, Tao Xu, Greg Brockman, Christine McLeavey, and Ilya Sutskever. 2023. Robust speech recognition via large-scale weak supervision. In International conference on machine learning. PMLR, 28492--28518."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2023.107013"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/EMC2-NIPS53020.2019.00021"},{"volume-title":"Spoken language understanding: Systems for extracting semantic information from speech","author":"Tur Gokhan","key":"e_1_3_2_1_14_1","unstructured":"Gokhan Tur and Renato De Mori. 2011. Spoken language understanding: Systems for extracting semantic information from speech. John Wiley & Sons."},{"key":"e_1_3_2_1_15_1","volume-title":"Sequence-to-sequence models can directly translate foreign speech. arXiv preprint arXiv:1703.08581","author":"Weiss Ron J","year":"2017","unstructured":"Ron J Weiss, Jan Chorowski, Navdeep Jaitly, Yonghui Wu, and Zhifeng Chen. 2017. Sequence-to-sequence models can directly translate foreign speech. arXiv preprint arXiv:1703.08581 (2017)."},{"key":"e_1_3_2_1_16_1","volume-title":"Paraphrasing with large language models. arXiv preprint arXiv:1911.09661","author":"Witteveen Sam","year":"2019","unstructured":"Sam Witteveen and Martin Andrews. 2019. Paraphrasing with large language models. arXiv preprint arXiv:1911.09661 (2019)."}],"event":{"name":"CIKM '25: The 34th ACM International Conference on Information and Knowledge Management","sponsor":["SIGIR ACM Special Interest Group on Information Retrieval","SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"],"location":"Seoul Republic of Korea","acronym":"CIKM '25"},"container-title":["Proceedings of the 34th ACM International Conference on Information and Knowledge Management"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3746252.3761477","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,12]],"date-time":"2025-12-12T00:03:24Z","timestamp":1765497804000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3746252.3761477"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,10]]},"references-count":16,"alternative-id":["10.1145\/3746252.3761477","10.1145\/3746252"],"URL":"https:\/\/doi.org\/10.1145\/3746252.3761477","relation":{},"subject":[],"published":{"date-parts":[[2025,11,10]]},"assertion":[{"value":"2025-11-10","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}