{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,30]],"date-time":"2025-12-30T16:59:35Z","timestamp":1767113975815,"version":"3.48.0"},"publisher-location":"New York, NY, USA","reference-count":24,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,10,12]]},"DOI":"10.1145\/3714394.3750571","type":"proceedings-article","created":{"date-parts":[[2025,12,29]],"date-time":"2025-12-29T21:13:49Z","timestamp":1767042829000},"page":"1697-1700","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Intelligence-enabled IoT with Gen AI, Powered by Model Compression and Optimization"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1794-6719","authenticated-orcid":false,"given":"Arijit","family":"Ukil","sequence":"first","affiliation":[{"name":"TCS Research, Tata Consultancy Services, Kolkata, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2651-6684","authenticated-orcid":false,"given":"Antonio J.","family":"Jara","sequence":"additional","affiliation":[{"name":"Libelium, Murcia, Spain"}]}],"member":"320","published-online":{"date-parts":[[2025,12,29]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2017.2765695"},{"key":"e_1_3_2_1_2_1","volume-title":"Forty-first International Conference on Machine Learning.","author":"Cho Minsik","year":"2024","unstructured":"Minsik Cho, Mohammad Rastegari, and Devang Naik. 2024. Kv-runahead: Scalable causal llm inference by parallel key-value cache generation. In Forty-first International Conference on Machine Learning."},{"key":"e_1_3_2_1_3_1","volume-title":"Qlora: Efficient finetuning of quantized llms. Advances in neural information processing systems 36","author":"Dettmers Tim","year":"2023","unstructured":"Tim Dettmers, Artidoro Pagnoni, Ari Holtzman, and Luke Zettlemoyer. 2023. Qlora: Efficient finetuning of quantized llms. Advances in neural information processing systems 36 (2023), 10088-10115."},{"key":"e_1_3_2_1_4_1","volume-title":"The lottery ticket hypothesis: Finding sparse, trainable neural networks. arXiv preprint arXiv:1803.03635","author":"Frankle Jonathan","year":"2018","unstructured":"Jonathan Frankle and Michael Carbin. 2018. The lottery ticket hypothesis: Finding sparse, trainable neural networks. arXiv preprint arXiv:1803.03635 (2018)."},{"key":"e_1_3_2_1_5_1","unstructured":"Daya Guo Dejian Yang Haowei Zhang Junxiao Song Ruoyu Zhang Runxin Xu Qihao Zhu Shirong Ma Peiyi Wang Xiao Bi et al. 2025. Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning. arXiv preprint arXiv:2501.12948 (2025)."},{"key":"e_1_3_2_1_6_1","volume-title":"Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531","author":"Hinton Geoffrey","year":"2015","unstructured":"Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. 2015. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015)."},{"key":"e_1_3_2_1_7_1","first-page":"3","article-title":"Lora: Low-rank adaptation of large language models","volume":"1","author":"Hu Edward J","year":"2022","unstructured":"Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, et al. 2022. Lora: Low-rank adaptation of large language models. ICLR 1, 2 (2022), 3.","journal-title":"ICLR"},{"key":"e_1_3_2_1_8_1","volume-title":"Tinybert: Distilling bert for natural language understanding. arXiv preprint arXiv:1909.10351","author":"Jiao Xiaoqi","year":"2019","unstructured":"Xiaoqi Jiao, Yichun Yin, Lifeng Shang, Xin Jiang, Xiao Chen, Linlin Li, FangWang, and Qun Liu. 2019. Tinybert: Distilling bert for natural language understanding. arXiv preprint arXiv:1909.10351 (2019)."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/EMBC48229.2022.9871259"},{"key":"e_1_3_2_1_10_1","volume-title":"a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108","author":"Sanh Victor","year":"2019","unstructured":"Victor Sanh, Lysandre Debut, Julien Chaumond, and Thomas Wolf. 2019. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108 (2019)."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICC.2008.427"},{"key":"e_1_3_2_1_12_1","volume-title":"Asimple and effective pruning approach for large language models. arXiv preprint arXiv:2306.11695","author":"Sun Mingjie","year":"2023","unstructured":"Mingjie Sun, Zhuang Liu, Anna Bair, and J Zico Kolter. 2023. Asimple and effective pruning approach for large language models. arXiv preprint arXiv:2306.11695 (2023)."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISCC.2014.6912486"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.3390\/s19122733"},{"key":"e_1_3_2_1_15_1","volume-title":"Deep Neural Network Model Compression and Signal Processing. In 2024 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW). IEEE, 179--183","author":"Ukil Arijit","year":"2024","unstructured":"Arijit Ukil, Angshul Majumdar, Antonio J Jara, and Jo\u00e3o Gama. 2024. Deep Neural Network Model Compression and Signal Processing. In 2024 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW). IEEE, 179--183."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0277975"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1109\/JSEN.2022.3162691"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"crossref","unstructured":"Arijit Ukil and Uttam Kumar Roy. 2017. Smart cardiac health management in IoT through heart sound signal analytics and robust noise filtering. In 2017 IEEE 28th annual international symposium on personal indoor and mobile radio communications (PIMRC). IEEE 1--5.","DOI":"10.1109\/PIMRC.2017.8292659"},{"volume-title":"Structured Lottery Ticket Hypothesis for Effective Deep Neural Network Model Size Reduction. In 2024 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW)","author":"Ukil Arijit","key":"e_1_3_2_1_19_1","unstructured":"Arijit Ukil, Ishan Sahu, Mridul Biswas, Arpan Pal, and Angshul Majumdar. 2024. Structured Lottery Ticket Hypothesis for Effective Deep Neural Network Model Size Reduction. In 2024 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW). IEEE, 330--334."},{"key":"e_1_3_2_1_20_1","volume-title":"Linformer: Self-attention with linear complexity. arXiv preprint arXiv:2006.04768","author":"Wang Sinong","year":"2020","unstructured":"Sinong Wang, Belinda Z Li, Madian Khabsa, Han Fang, and Hao Ma. 2020. Linformer: Self-attention with linear complexity. arXiv preprint arXiv:2006.04768 (2020)."},{"key":"e_1_3_2_1_21_1","volume-title":"International Conference on Machine Learning. PMLR, 38087-38099","author":"Xiao Guangxuan","year":"2023","unstructured":"Guangxuan Xiao, Ji Lin, Mickael Seznec, Hao Wu, Julien Demouth, and Song Han. 2023. Smoothquant: Accurate and efficient post-training quantization for large language models. In International Conference on Machine Learning. PMLR, 38087-38099."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3625687.3625793"},{"key":"e_1_3_2_1_23_1","volume-title":"Forty-first International Conference on Machine Learning.","author":"Yin Lu","year":"2024","unstructured":"Lu Yin, AJAY KUMAR JAISWAL, Shiwei Liu, Souvik Kundu, and Zhangyang Wang. 2024. Junk DNA Hypothesis: Pruning Small Pre-Trained Weights Irreversibly and Monotonically Impairs''Difficult'' Downstream Tasks in LLMs. In Forty-first International Conference on Machine Learning."},{"key":"e_1_3_2_1_24_1","volume-title":"Long-short transformer: Efficient transformers for language and vision. Advances in neural information processing systems 34","author":"Zhu Chen","year":"2021","unstructured":"Chen Zhu, Wei Ping, Chaowei Xiao, Mohammad Shoeybi, Tom Goldstein, Anima Anandkumar, and Bryan Catanzaro. 2021. Long-short transformer: Efficient transformers for language and vision. Advances in neural information processing systems 34 (2021), 17723--17736."}],"event":{"name":"UbiComp '25:The 2025 ACM International Joint Conference on Pervasive and Ubiquitous Computing \/ ISWC ACM International Symposium on Wearable Computers","sponsor":["SIGMOBILE ACM Special Interest Group on Mobility of Systems, Users, Data and Computing","SIGCHI ACM Special Interest Group on Computer-Human Interaction","SIGSPATIAL ACM Special Interest Group on Spatial Information"],"location":"Espoo Finland"},"container-title":["Companion of the 2025 ACM International Joint Conference on Pervasive and Ubiquitous Computing"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3714394.3750571","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,30]],"date-time":"2025-12-30T16:55:15Z","timestamp":1767113715000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3714394.3750571"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,12]]},"references-count":24,"alternative-id":["10.1145\/3714394.3750571","10.1145\/3714394"],"URL":"https:\/\/doi.org\/10.1145\/3714394.3750571","relation":{},"subject":[],"published":{"date-parts":[[2025,10,12]]},"assertion":[{"value":"2025-12-29","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}