{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,15]],"date-time":"2025-11-15T07:13:35Z","timestamp":1763190815143,"version":"3.45.0"},"reference-count":32,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,6,30]],"date-time":"2025-06-30T00:00:00Z","timestamp":1751241600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,6,30]],"date-time":"2025-06-30T00:00:00Z","timestamp":1751241600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100006190","name":"Research and Development","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100006190","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,6,30]]},"DOI":"10.1109\/ijcnn64981.2025.11227876","type":"proceedings-article","created":{"date-parts":[[2025,11,14]],"date-time":"2025-11-14T18:46:15Z","timestamp":1763145975000},"page":"1-8","source":"Crossref","is-referenced-by-count":0,"title":["FedSIT: Efficient Federated Fine-Tuning with Model Splitting and Importance-Based Tuning"],"prefix":"10.1109","author":[{"given":"Xinyu","family":"Gu","sequence":"first","affiliation":[{"name":"Shenzhen Campus of Sun Yat-Sen University,School of Electronics and Communication Engineering,Shenzhen,China"}]},{"given":"Xinghua","family":"Sun","sequence":"additional","affiliation":[{"name":"Shenzhen Campus of Sun Yat-Sen University,School of Electronics and Communication Engineering,Shenzhen,China"}]},{"given":"Chenyuan","family":"Feng","sequence":"additional","affiliation":[{"name":"EURECOM,France"}]},{"given":"Xijun","family":"Wang","sequence":"additional","affiliation":[{"name":"Shenzhen Campus of Sun Yat-Sen University,School of Electronics and Communication Engineering,Shenzhen,China"}]},{"given":"Xiang","family":"Chen","sequence":"additional","affiliation":[{"name":"Sun Yat-Sen University,School of Electronics and Information Technology,Guangzhou,China"}]}],"member":"263","reference":[{"article-title":"Gpt-4 technical report","year":"2023","author":"Achiam","key":"ref1"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6239"},{"article-title":"Language models are few-shot learners","year":"2020","author":"B Brown","key":"ref3"},{"key":"ref4","first-page":"arXiv","article-title":"Federated large language model: A position paper","author":"Chen","year":"2023"},{"article-title":"Compressing large language models by streamlining the unimportant layer","year":"2024","author":"Chen","key":"ref5"},{"article-title":"Think you have solved question answering? try arc, the ai2 reasoning challenge","year":"2018","author":"Clark","key":"ref6"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-023-00626-4"},{"article-title":"Fate-llm: A industrial grade federated learning framework for large language models","year":"2023","author":"Fan","key":"ref8"},{"journal-title":"A framework for few-shot language model evaluation","year":"2024","author":"Gao","key":"ref9"},{"article-title":"The unreasonable ineffectiveness of the deeper layers","year":"2024","author":"Gromov","key":"ref10"},{"article-title":"What matters in transformers? not all attention is needed","year":"2024","author":"He","key":"ref11"},{"key":"ref12","first-page":"2790","article-title":"Parameter-efficient transfer learning for nlp","volume-title":"International conference on machine learning","author":"Houlsby"},{"article-title":"Lora: Low-rank adaptation of large language models","year":"2021","author":"Hu","key":"ref13"},{"key":"ref14","first-page":"1022","article-title":"Compacter: Efficient low-rank hypercomplex adapter layers","volume":"34","author":"Mahabadi","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref15","first-page":"2","article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","volume-title":"Proceedings of naacL-HLT","volume":"1","author":"Ming-Wei Chang Kenton"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1145\/3637528.3671573"},{"article-title":"Splitlora: A split parameter-efficient fine-tuning framework for large language models","year":"2024","author":"Lin","key":"ref17"},{"key":"ref18","first-page":"1273","article-title":"Communication-efficient learning of deep networks from decentralized data","author":"McMahan","year":"2017","journal-title":"Artificial intelligence and statistics"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2025.findings-acl.1035"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1260"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1145\/3460120.3485259"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.eacl-main.39"},{"journal-title":"Improving language understanding by generative pre-training","year":"2018","author":"Radford","key":"ref23"},{"article-title":"Llama 2: Open foundation and fine-tuned chat models","year":"2023","author":"Touvron","key":"ref24"},{"key":"ref25","article-title":"Attention is all you need","author":"Vaswani","year":"2017","journal-title":"Advances in Neural Information Processing Systems"},{"article-title":"Will we run out of data? limits of llm scaling based on human-generated data","year":"2024","author":"Villalobos","key":"ref26"},{"article-title":"Flora: Federated fine-tuning large language models with heterogeneous low-rank adaptations","year":"2024","author":"Wang","key":"ref27"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W17-4413"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1472"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP48485.2024.10447454"},{"article-title":"Adalora: Adaptive budget allocation for parameter-efficient fine-tuning","year":"2023","author":"Zhang","key":"ref32"}],"event":{"name":"2025 International Joint Conference on Neural Networks (IJCNN)","start":{"date-parts":[[2025,6,30]]},"location":"Rome, Italy","end":{"date-parts":[[2025,7,5]]}},"container-title":["2025 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11227166\/11227148\/11227876.pdf?arnumber=11227876","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,15]],"date-time":"2025-11-15T07:11:12Z","timestamp":1763190672000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11227876\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,30]]},"references-count":32,"URL":"https:\/\/doi.org\/10.1109\/ijcnn64981.2025.11227876","relation":{},"subject":[],"published":{"date-parts":[[2025,6,30]]}}}