{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,20]],"date-time":"2026-03-20T08:58:13Z","timestamp":1773997093517,"version":"3.50.1"},"reference-count":28,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,12,8]],"date-time":"2025-12-08T00:00:00Z","timestamp":1765152000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,12,8]],"date-time":"2025-12-08T00:00:00Z","timestamp":1765152000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,12,8]]},"DOI":"10.1109\/globecom59602.2025.11432128","type":"proceedings-article","created":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T20:04:01Z","timestamp":1773950641000},"page":"1035-1040","source":"Crossref","is-referenced-by-count":0,"title":["Task-Aware Parameter-Efficient Fine-Tuning of Large Pre-Trained Models at the Edge"],"prefix":"10.1109","author":[{"given":"Senkang","family":"Hu","sequence":"first","affiliation":[{"name":"Hong Kong JC STEM Lab of Smart City"}]},{"given":"Yanan","family":"Ma","sequence":"additional","affiliation":[{"name":"Hong Kong JC STEM Lab of Smart City"}]},{"given":"Yihang","family":"Tao","sequence":"additional","affiliation":[{"name":"Hong Kong JC STEM Lab of Smart City"}]},{"given":"Zhengru","family":"Fang","sequence":"additional","affiliation":[{"name":"Hong Kong JC STEM Lab of Smart City"}]},{"given":"Zihan","family":"Fang","sequence":"additional","affiliation":[{"name":"Hong Kong JC STEM Lab of Smart City"}]},{"given":"Yiqin","family":"Deng","sequence":"additional","affiliation":[{"name":"Hong Kong JC STEM Lab of Smart City"}]},{"given":"Sam Tak","family":"Wu Kwong","sequence":"additional","affiliation":[{"name":"Lingnan University,Department of Computing and Decision Sciences"}]},{"given":"Yuguang","family":"Fang","sequence":"additional","affiliation":[{"name":"Hong Kong JC STEM Lab of Smart City"}]}],"member":"263","reference":[{"key":"ref1","article-title":"DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning","author":"Guo","year":"2025"},{"key":"ref2","article-title":"Evaluation of OpenAI o1: Opportunities and Challenges of AGI","author":"Zhong","year":"2024"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.01432"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i3.27963"},{"key":"ref5","article-title":"GaLore: memory-efficient LLM training by gradient low-rank projection","volume-title":"Proceedings of the 41st International Conference on Machine Learning","author":"Zhao"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.378"},{"key":"ref7","article-title":"LoRA: Low-Rank Adaptation of Large Language Models","volume-title":"International Conference on Learning Representations","author":"Hu"},{"key":"ref8","article-title":"Chain of LoRA: Efficient Fine-tuning of Language Models via Residual Learning","author":"Xia","year":"2024"},{"key":"ref9","article-title":"ReLoRA: High-Rank Training Through Low-Rank Updates","volume-title":"The Twelfth International Conference on Learning Representations","author":"Lialin"},{"key":"ref10","article-title":"Parameter-Efficient Fine-Tuning for Large Models: A Comprehensive Survey","author":"Han","year":"2024"},{"key":"ref11","first-page":"2790","article-title":"Parameter-Efficient Transfer Learning for NLP","volume-title":"Proceedings of the 36th International Conference on Machine Learning","volume":"97","author":"Houlsby"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01086"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00432"},{"key":"ref14","first-page":"10 842","article-title":"Data Valuation using Reinforcement Learning","volume-title":"Proceedings of the 37th International Conference on Machine Learning","volume":"119","author":"Yoon"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00148"},{"key":"ref16","first-page":"7294","article-title":"LEEP: A New Measure to Evaluate Transferability of Learned Representations","volume-title":"Proceedings of the 37th International Conference on Machine Learning","volume":"119","author":"Nguyen"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.02699"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i11.26505"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2025.3527641"},{"key":"ref20","article-title":"LLM.int8(): 8-bit matrix multiplication for transformers at scale","volume-title":"Proceedings of the 36th International Conference on Neural Information Processing Systems","author":"Dettmers"},{"key":"ref21","article-title":"SparseGPT: massive language models can be accurately pruned in one-shot","volume-title":"Proceedings of the 40th International Conference on Machine Learning","author":"Frantar"},{"key":"ref22","article-title":"MINILM: deep self-attention distillation for task-agnostic compression of pre-trained transformers","volume-title":"Proceedings of the 34th International Conference on Neural Information Processing Systems","author":"Wang"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19827-4_41"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-short.1"},{"key":"ref25","article-title":"LORA: LOW-RANK ADAPTATION OF LARGE LANGUAGE MODELS","author":"Hu","year":"2022"},{"key":"ref26","article-title":"A Large-scale Study of Representation Learning with the Visual Task Adaptation Benchmark","author":"Zhai","year":"2020"},{"key":"ref27","article-title":"An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale","volume-title":"International Conference on Learning Representations","author":"Dosovitskiy"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"}],"event":{"name":"GLOBECOM 2025 - 2025 IEEE Global Communications Conference","location":"Taipei, Taiwan","start":{"date-parts":[[2025,12,8]]},"end":{"date-parts":[[2025,12,12]]}},"container-title":["GLOBECOM 2025 - 2025 IEEE Global Communications Conference"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11431620\/11431622\/11432128.pdf?arnumber=11432128","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,20]],"date-time":"2026-03-20T05:29:35Z","timestamp":1773984575000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11432128\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,8]]},"references-count":28,"URL":"https:\/\/doi.org\/10.1109\/globecom59602.2025.11432128","relation":{},"subject":[],"published":{"date-parts":[[2025,12,8]]}}}