{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,30]],"date-time":"2026-03-30T21:05:28Z","timestamp":1774904728854,"version":"3.50.1"},"reference-count":48,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"7","license":[{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Guangdong S&T Programme","award":["2024B0101030002"],"award-info":[{"award-number":["2024B0101030002"]}]},{"name":"Basic Research Project of Hetao Shenzhen-HK S&T Cooperation Zone","award":["HZQB-KCZYZ-2021067"],"award-info":[{"award-number":["HZQB-KCZYZ-2021067"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62501514"],"award-info":[{"award-number":["62501514"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62293482"],"award-info":[{"award-number":["62293482"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Shenzhen Outstanding Talents Training Fund","award":["202002"],"award-info":[{"award-number":["202002"]}]},{"name":"Guangdong Research Projects","award":["2017ZT07X152"],"award-info":[{"award-number":["2017ZT07X152"]}]},{"name":"Guangdong Research Projects","award":["2019CX01X104"],"award-info":[{"award-number":["2019CX01X104"]}]},{"name":"Guangdong Provincial Key Laboratory of Future Networks of Intelligence","award":["2022B1212010001"],"award-info":[{"award-number":["2022B1212010001"]}]},{"name":"Shenzhen Key Laboratory of Big Data and Artificial Intelligence","award":["SYSPG20241211173853027"],"award-info":[{"award-number":["SYSPG20241211173853027"]}]},{"DOI":"10.13039\/100017942","name":"Guangdong Province Radio Science Data Center","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100017942","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Internet Things J."],"published-print":{"date-parts":[[2026,4,1]]},"DOI":"10.1109\/jiot.2026.3654102","type":"journal-article","created":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T21:00:27Z","timestamp":1769202027000},"page":"14882-14897","source":"Crossref","is-referenced-by-count":1,"title":["Adaptive Pruning for Large Language Models With Structural Importance Awareness"],"prefix":"10.1109","volume":"13","author":[{"given":"Haotian","family":"Zheng","sequence":"first","affiliation":[{"name":"Shenzhen Future Network of Intelligence Institute (FNii-Shenzhen), The Chinese University of Hong Kong, Shenzhen, Guangdong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7409-6127","authenticated-orcid":false,"given":"Jinke","family":"Ren","sequence":"additional","affiliation":[{"name":"School of Science and Engineering, FNii-Shenzhen, and Guangdong Provincial Key Laboratory of Future Networks of Intelligence, The Chinese University of Hong Kong, Shenzhen, Guangdong, China"}]},{"given":"Yatong","family":"Han","sequence":"additional","affiliation":[{"name":"FNii-Shenzhen, The Chinese University of Hong Kong, Shenzhen, Guangdong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5312-2413","authenticated-orcid":false,"given":"Yushan","family":"Sun","sequence":"additional","affiliation":[{"name":"National Key Laboratory of Autonomous Marine Vehicle Technology, Harbin Engineering University, Harbin, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6859-3645","authenticated-orcid":false,"given":"Ruichen","family":"Zhang","sequence":"additional","affiliation":[{"name":"College of Computing and Data Science, Nanyang Technological University, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-0566-0749","authenticated-orcid":false,"given":"Wenbo","family":"Zhang","sequence":"additional","affiliation":[{"name":"Hangzhou RoboCT Technology Development Company Ltd., Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7669-2686","authenticated-orcid":false,"given":"Zhen","family":"Li","sequence":"additional","affiliation":[{"name":"School of Science and Engineering, FNii-Shenzhen, and Guangdong Provincial Key Laboratory of Future Networks of Intelligence, The Chinese University of Hong Kong, Shenzhen, Guangdong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7442-7416","authenticated-orcid":false,"given":"Dusit","family":"Niyato","sequence":"additional","affiliation":[{"name":"College of Computing and Data Science, Nanyang Technological University, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2608-775X","authenticated-orcid":false,"given":"Shuguang","family":"Cui","sequence":"additional","affiliation":[{"name":"School of Science and Engineering, FNii-Shenzhen, and Guangdong Provincial Key Laboratory of Future Networks of Intelligence, The Chinese University of Hong Kong, Shenzhen, Guangdong, China"}]}],"member":"263","reference":[{"key":"ref1","article-title":"A survey of large language models","author":"Zhao","year":"2023","journal-title":"arXiv:2303.18223"},{"key":"ref2","article-title":"Emergent abilities of large language models","author":"Wei","year":"2022","journal-title":"Trans. Mach. Learn. Res."},{"key":"ref3","article-title":"A survey on large language models: Applications, challenges, limitations, and practical usage","author":"Hadi","year":"2023","journal-title":"TechRxiv"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/BigData62323.2024.10825187"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/MIC.2024.3443169"},{"key":"ref6","article-title":"Sparks of artificial general intelligence: Early experiments with GPT-4","author":"Bubeck","year":"2023","journal-title":"arXiv:2303.12712"},{"key":"ref7","article-title":"Deep compression: Compressing deep neural networks with pruning, trained quantization and Huffman coding","volume-title":"Proc. Int. Conf. Learn. Represent. (ICLR)","author":"Han"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2024.3447085"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00704"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.findings-acl.178"},{"key":"ref11","article-title":"A simple and effective pruning approach for large language models","volume-title":"Proc. Int. Conf. Learn. Represent. (ICLR)","author":"Sun"},{"key":"ref12","first-page":"21702","article-title":"LLM-pruner: On the structural pruning of large language models","volume-title":"Proc. Adv. Neural Inf. Process. Syst. (NeurIPS)","author":"Ma"},{"key":"ref13","article-title":"LoRAShear: Efficient large language model structured pruning and knowledge recovery","author":"Chen","year":"2023","journal-title":"arXiv:2310.18356"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i10.28960"},{"key":"ref15","article-title":"Dynamic sparse graph for efficient deep learning","volume-title":"Proc. Int. Conf. Learn. Represent. (ICLR)","author":"Liu"},{"key":"ref16","first-page":"65597","article-title":"ZipLM: Inference-aware structured pruning of language models","volume-title":"Proc. Adv. Neural Inf. Process. Syst. (NeurIPS)","author":"Kurtic"},{"key":"ref17","article-title":"GPT-4 technical report","volume-title":"arXiv:2303.08774","author":"Achiam","year":"2023"},{"key":"ref18","article-title":"LLaMA: Open and efficient foundation language models","author":"Touvron","year":"2023","journal-title":"arXiv:2302.13971"},{"key":"ref19","volume-title":"Vicuna: An Open-Source Chatbot Impressing GPT-4 With 90%* ChatGPT Quality","author":"Chiang","year":"2023"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.26"},{"key":"ref21","first-page":"10323","article-title":"SparseGPT: Massive language models can be accurately pruned in one-shot","volume-title":"Proc. Int. Conf. Mach. Learn. (ICML)","author":"Frantar"},{"key":"ref22","article-title":"Pruning large language models via accuracy predictor","author":"Ji","year":"2023","journal-title":"arXiv:2309.09507"},{"key":"ref23","article-title":"Shortened LLaMA: A simple depth pruning for large language models","volume-title":"Proc. Int. Conf. Learn. Represent. (ICLR)","author":"Kim"},{"key":"ref24","article-title":"MINI-LLM: Memory-efficient structured pruning for large language models","author":"Cheng","year":"2024","journal-title":"arXiv:2407.11681"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.52202\/079017-0289"},{"key":"ref26","article-title":"Parameter-efficient fine-tuning for large models: A comprehensive survey","author":"Han","year":"2024","journal-title":"Trans. Mach. Learn. Res."},{"key":"ref27","first-page":"2790","article-title":"Parameter-efficient transfer learning for NLP","volume-title":"Proc. Int. Conf. Mach. Learn. (ICML)","author":"Houlsby"},{"key":"ref28","article-title":"LoRA: Low-rank adaptation of large language models","volume-title":"Proc. Int. Conf. Learn. Represent. (ICLR)","author":"Hu"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.52202\/075280-0441"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1145\/3649329.3658473"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2024.3524255"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/TMC.2024.3513457"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01544"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2024.3416316"},{"key":"ref35","article-title":"Llama 2: Open foundation and fine-tuned chat models","author":"Touvron","year":"2023","journal-title":"arXiv:2307.09288"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1007\/s44267-024-00070-x"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.21236\/ADA273556"},{"key":"ref38","article-title":"Pointer sentinel mixture models","volume-title":"Proc. Int. Conf. Learn. Represent. (ICLR)","author":"Merity"},{"key":"ref39","article-title":"Think you have solved question answering? Try ARC, the AI2 reasoning challenge","author":"Clark","year":"2018","journal-title":"arXiv:1803.05457"},{"key":"ref40","first-page":"2924","article-title":"BoolQ: Exploring the surprising difficulty of natural yes\/no questions","volume-title":"Proc. Conf. North Amer. Chapter Assoc. Comput. Linguistics, Hum. Lang. Technol. (NAACL-HLT)","author":"Clark"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/p19-1472"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6239"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6399"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1260"},{"key":"ref45","article-title":"Reassessing layer pruning in LLMs: New insights and methods","author":"Lu","year":"2024","journal-title":"arXiv:2411.15558"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.11"},{"key":"ref47","article-title":"WizardLM: Empowering large pre-trained language models to follow complex instructions","author":"Xu","year":"2023","journal-title":"arXiv:2304.12244"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/ITEC60657.2024.10598994"}],"container-title":["IEEE Internet of Things Journal"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6488907\/11456203\/11360603.pdf?arnumber=11360603","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,30]],"date-time":"2026-03-30T20:08:12Z","timestamp":1774901292000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11360603\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,4,1]]},"references-count":48,"journal-issue":{"issue":"7"},"URL":"https:\/\/doi.org\/10.1109\/jiot.2026.3654102","relation":{},"ISSN":["2327-4662","2372-2541"],"issn-type":[{"value":"2327-4662","type":"electronic"},{"value":"2372-2541","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,4,1]]}}}